[ 475.541435] env[61857]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61857) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 475.541770] env[61857]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61857) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 475.541873] env[61857]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61857) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 475.542268] env[61857]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 475.638763] env[61857]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61857) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 475.647905] env[61857]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.009s {{(pid=61857) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 476.249575] env[61857]: INFO nova.virt.driver [None req-50a8c15b-7332-4016-8d10-1cc5e4924f7e None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 476.320726] env[61857]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 476.320911] env[61857]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 476.321046] env[61857]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61857) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 479.410942] env[61857]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-6501f674-e44b-425a-bd2e-03d2f3cbcb2d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 479.426756] env[61857]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61857) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 479.426971] env[61857]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-9199c938-12ab-45c1-b79d-fe59ea7af7d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 479.469303] env[61857]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 96e31. [ 479.469449] env[61857]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.149s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 479.469953] env[61857]: INFO nova.virt.vmwareapi.driver [None req-50a8c15b-7332-4016-8d10-1cc5e4924f7e None None] VMware vCenter version: 7.0.3 [ 479.473325] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a58db3d-d591-4741-ac02-c4f94c3d65f7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 479.490133] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4074f8ba-47f0-4d2f-898b-7a0724fd1016 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 479.495715] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c460086-6ed8-4b1c-9ed8-842cd631419d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 479.502076] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8090112-bb19-4572-a5fb-b21de560a918 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 479.515147] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50762fed-84a6-4606-aa05-e4bbd4e74510 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 479.520787] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24b412e-db42-4e08-9d7d-aaf627a3f718 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 479.550076] env[61857]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-32db6b11-ce2c-43f8-9946-aeb60786146d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 479.554997] env[61857]: DEBUG nova.virt.vmwareapi.driver [None req-50a8c15b-7332-4016-8d10-1cc5e4924f7e None None] Extension org.openstack.compute already exists. {{(pid=61857) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 479.557623] env[61857]: INFO nova.compute.provider_config [None req-50a8c15b-7332-4016-8d10-1cc5e4924f7e None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 480.060765] env[61857]: DEBUG nova.context [None req-50a8c15b-7332-4016-8d10-1cc5e4924f7e None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),e2829645-077f-46e3-8005-83bb424858ce(cell1) {{(pid=61857) load_cells /opt/stack/nova/nova/context.py:464}} [ 480.063200] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 480.063431] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 480.064111] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 480.064546] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Acquiring lock "e2829645-077f-46e3-8005-83bb424858ce" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 480.064764] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Lock "e2829645-077f-46e3-8005-83bb424858ce" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 480.069038] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Lock "e2829645-077f-46e3-8005-83bb424858ce" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 480.086146] env[61857]: INFO dbcounter [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Registered counter for database nova_cell0 [ 480.094248] env[61857]: INFO dbcounter [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Registered counter for database nova_cell1 [ 480.097548] env[61857]: DEBUG oslo_db.sqlalchemy.engines [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61857) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 480.097901] env[61857]: DEBUG oslo_db.sqlalchemy.engines [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61857) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 480.102767] env[61857]: ERROR nova.db.main.api [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 480.102767] env[61857]: result = function(*args, **kwargs) [ 480.102767] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 480.102767] env[61857]: return func(*args, **kwargs) [ 480.102767] env[61857]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 480.102767] env[61857]: result = fn(*args, **kwargs) [ 480.102767] env[61857]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 480.102767] env[61857]: return f(*args, **kwargs) [ 480.102767] env[61857]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 480.102767] env[61857]: return db.service_get_minimum_version(context, binaries) [ 480.102767] env[61857]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 480.102767] env[61857]: _check_db_access() [ 480.102767] env[61857]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 480.102767] env[61857]: stacktrace = ''.join(traceback.format_stack()) [ 480.102767] env[61857]: [ 480.103588] env[61857]: ERROR nova.db.main.api [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 480.103588] env[61857]: result = function(*args, **kwargs) [ 480.103588] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 480.103588] env[61857]: return func(*args, **kwargs) [ 480.103588] env[61857]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 480.103588] env[61857]: result = fn(*args, **kwargs) [ 480.103588] env[61857]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 480.103588] env[61857]: return f(*args, **kwargs) [ 480.103588] env[61857]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 480.103588] env[61857]: return db.service_get_minimum_version(context, binaries) [ 480.103588] env[61857]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 480.103588] env[61857]: _check_db_access() [ 480.103588] env[61857]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 480.103588] env[61857]: stacktrace = ''.join(traceback.format_stack()) [ 480.103588] env[61857]: [ 480.104169] env[61857]: WARNING nova.objects.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Failed to get minimum service version for cell e2829645-077f-46e3-8005-83bb424858ce [ 480.104169] env[61857]: WARNING nova.objects.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 480.104518] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Acquiring lock "singleton_lock" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 480.104677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Acquired lock "singleton_lock" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 480.104918] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Releasing lock "singleton_lock" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 480.105256] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Full set of CONF: {{(pid=61857) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 480.105401] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ******************************************************************************** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 480.105528] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Configuration options gathered from: {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 480.105662] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 480.105850] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 480.105976] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ================================================================================ {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 480.106197] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] allow_resize_to_same_host = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.106370] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] arq_binding_timeout = 300 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.106500] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] backdoor_port = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.106623] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] backdoor_socket = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.106784] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] block_device_allocate_retries = 60 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.106945] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] block_device_allocate_retries_interval = 3 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.107125] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cert = self.pem {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.107292] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.107459] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute_monitors = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.107622] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] config_dir = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.107788] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] config_drive_format = iso9660 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.107921] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.108115] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] config_source = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.108301] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] console_host = devstack {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.108468] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] control_exchange = nova {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.108627] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cpu_allocation_ratio = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.108785] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] daemon = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.108950] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] debug = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.109124] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] default_access_ip_network_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.109291] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] default_availability_zone = nova {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.109445] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] default_ephemeral_format = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.109601] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] default_green_pool_size = 1000 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.109831] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.109993] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] default_schedule_zone = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.110195] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] disk_allocation_ratio = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.110361] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] enable_new_services = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.110538] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] enabled_apis = ['osapi_compute'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.110795] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] enabled_ssl_apis = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.110964] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] flat_injected = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.111157] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] force_config_drive = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.111326] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] force_raw_images = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.111496] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] graceful_shutdown_timeout = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.111655] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] heal_instance_info_cache_interval = 60 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.111870] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] host = cpu-1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.112058] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.112230] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.112393] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.112616] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.112781] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] instance_build_timeout = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.112944] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] instance_delete_interval = 300 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.113130] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] instance_format = [instance: %(uuid)s] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.113301] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] instance_name_template = instance-%08x {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.113463] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] instance_usage_audit = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.113635] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] instance_usage_audit_period = month {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.113803] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.113970] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.114151] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] internal_service_availability_zone = internal {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.114312] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] key = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.114471] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] live_migration_retry_count = 30 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.114639] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] log_color = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.114803] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] log_config_append = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.114967] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.115138] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] log_dir = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.115298] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] log_file = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.115425] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] log_options = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.115585] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] log_rotate_interval = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.115749] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] log_rotate_interval_type = days {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.115913] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] log_rotation_type = none {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.116050] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.116177] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.116344] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.116508] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.116634] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.116793] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] long_rpc_timeout = 1800 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.116949] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] max_concurrent_builds = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.117117] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] max_concurrent_live_migrations = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.117276] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] max_concurrent_snapshots = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.117431] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] max_local_block_devices = 3 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.117586] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] max_logfile_count = 30 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.117739] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] max_logfile_size_mb = 200 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.117893] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] maximum_instance_delete_attempts = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.118088] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] metadata_listen = 0.0.0.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.118273] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] metadata_listen_port = 8775 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.118442] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] metadata_workers = 2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.118604] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] migrate_max_retries = -1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.118773] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] mkisofs_cmd = genisoimage {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.118981] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.119127] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] my_ip = 10.180.1.21 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.119294] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] network_allocate_retries = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.119474] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.119639] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.119801] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] osapi_compute_listen_port = 8774 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.119970] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] osapi_compute_unique_server_name_scope = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.120173] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] osapi_compute_workers = 2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.120347] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] password_length = 12 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.120509] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] periodic_enable = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.120739] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] periodic_fuzzy_delay = 60 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.120937] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] pointer_model = usbtablet {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.121141] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] preallocate_images = none {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.121315] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] publish_errors = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.121447] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] pybasedir = /opt/stack/nova {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.121607] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ram_allocation_ratio = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.121768] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] rate_limit_burst = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.121936] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] rate_limit_except_level = CRITICAL {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.122206] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] rate_limit_interval = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.122374] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] reboot_timeout = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.122534] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] reclaim_instance_interval = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.122691] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] record = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.122859] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] reimage_timeout_per_gb = 60 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.123036] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] report_interval = 120 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.123205] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] rescue_timeout = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.123366] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] reserved_host_cpus = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.123524] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] reserved_host_disk_mb = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.123684] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] reserved_host_memory_mb = 512 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.123845] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] reserved_huge_pages = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.124009] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] resize_confirm_window = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.124177] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] resize_fs_using_block_device = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.124338] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] resume_guests_state_on_host_boot = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.124504] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.124667] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] rpc_response_timeout = 60 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.124826] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] run_external_periodic_tasks = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.124992] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] running_deleted_instance_action = reap {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.125166] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.125325] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] running_deleted_instance_timeout = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.125483] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler_instance_sync_interval = 120 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.125649] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_down_time = 720 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.125817] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] servicegroup_driver = db {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.125972] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] shell_completion = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.126146] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] shelved_offload_time = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.126307] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] shelved_poll_interval = 3600 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.126474] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] shutdown_timeout = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.126633] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] source_is_ipv6 = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.126794] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ssl_only = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.127047] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.127222] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] sync_power_state_interval = 600 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.127386] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] sync_power_state_pool_size = 1000 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.127553] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] syslog_log_facility = LOG_USER {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.127711] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] tempdir = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.127871] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] timeout_nbd = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.128063] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] transport_url = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.128248] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] update_resources_interval = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.128414] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] use_cow_images = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.128576] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] use_eventlog = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.128734] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] use_journal = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.128893] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] use_json = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.129074] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] use_rootwrap_daemon = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.129230] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] use_stderr = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.129388] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] use_syslog = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.129543] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vcpu_pin_set = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.129709] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plugging_is_fatal = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.129874] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plugging_timeout = 300 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.130064] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] virt_mkfs = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.130245] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] volume_usage_poll_interval = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.130409] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] watch_log_file = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.130579] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] web = /usr/share/spice-html5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 480.130825] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_concurrency.disable_process_locking = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.131426] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.131623] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.131799] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.131976] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.132166] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.132338] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.132521] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.auth_strategy = keystone {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.132687] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.compute_link_prefix = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.132861] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.133042] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.dhcp_domain = novalocal {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.133219] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.enable_instance_password = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.133385] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.glance_link_prefix = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.133551] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.133721] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.133894] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.instance_list_per_project_cells = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.134072] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.list_records_by_skipping_down_cells = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.134244] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.local_metadata_per_cell = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.134481] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.max_limit = 1000 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.134668] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.metadata_cache_expiration = 15 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.134847] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.neutron_default_tenant_id = default {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.135031] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.response_validation = warn {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.135211] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.use_neutron_default_nets = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.135381] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.135544] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.135714] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.135885] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.136065] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.vendordata_dynamic_targets = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.136235] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.vendordata_jsonfile_path = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.136416] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.136607] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.backend = dogpile.cache.memcached {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.136777] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.backend_argument = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.136952] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.config_prefix = cache.oslo {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.137140] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.dead_timeout = 60.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.137310] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.debug_cache_backend = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.137474] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.enable_retry_client = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.137639] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.enable_socket_keepalive = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.137811] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.enabled = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.137977] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.enforce_fips_mode = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.138178] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.expiration_time = 600 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.138350] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.hashclient_retry_attempts = 2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.138519] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.138683] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.memcache_dead_retry = 300 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.138843] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.memcache_password = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.139013] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.139189] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.139354] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.memcache_pool_maxsize = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.139516] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.139679] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.memcache_sasl_enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.139859] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.140052] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.140246] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.memcache_username = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.140421] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.proxies = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.140587] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.redis_db = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.140749] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.redis_password = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.141013] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.141227] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.141409] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.redis_server = localhost:6379 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.141579] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.redis_socket_timeout = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.141741] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.redis_username = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.141908] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.retry_attempts = 2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.142091] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.retry_delay = 0.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.142265] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.socket_keepalive_count = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.142430] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.socket_keepalive_idle = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.142593] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.socket_keepalive_interval = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.142752] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.tls_allowed_ciphers = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.142911] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.tls_cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.143082] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.tls_certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.143248] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.tls_enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.143409] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cache.tls_keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.143582] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.auth_section = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.143755] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.auth_type = password {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.143920] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.144108] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.144276] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.144440] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.144607] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.cross_az_attach = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.144769] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.debug = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.144929] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.endpoint_template = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.145107] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.http_retries = 3 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.145278] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.145440] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.145612] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.os_region_name = RegionOne {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.145777] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.145938] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cinder.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.146126] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.146290] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.cpu_dedicated_set = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.146450] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.cpu_shared_set = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.146617] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.image_type_exclude_list = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.146783] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.146947] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.147123] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.147291] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.147458] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.147622] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.resource_provider_association_refresh = 300 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.147785] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.147947] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.shutdown_retry_interval = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.148167] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.148360] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] conductor.workers = 2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.148542] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] console.allowed_origins = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.148706] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] console.ssl_ciphers = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.148878] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] console.ssl_minimum_version = default {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.149062] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] consoleauth.enforce_session_timeout = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.149235] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] consoleauth.token_ttl = 600 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.149403] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.149561] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.149726] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.149889] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.connect_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.150081] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.connect_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.150260] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.endpoint_override = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.150428] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.150589] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.150751] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.max_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.150910] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.min_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.151201] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.region_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.151381] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.retriable_status_codes = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.151546] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.service_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.151718] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.service_type = accelerator {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.151883] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.152058] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.status_code_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.152226] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.status_code_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.152388] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.152567] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.152730] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] cyborg.version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.152909] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.backend = sqlalchemy {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.153093] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.connection = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.153265] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.connection_debug = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.153436] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.connection_parameters = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.153603] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.connection_recycle_time = 3600 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.153768] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.connection_trace = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.153931] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.db_inc_retry_interval = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.154107] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.db_max_retries = 20 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.154275] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.db_max_retry_interval = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.154439] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.db_retry_interval = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.154601] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.max_overflow = 50 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.154762] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.max_pool_size = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.154925] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.max_retries = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.155107] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.155270] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.mysql_wsrep_sync_wait = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.155429] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.pool_timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.155590] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.retry_interval = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.155750] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.slave_connection = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.155912] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.sqlite_synchronous = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.156084] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] database.use_db_reconnect = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.156267] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.backend = sqlalchemy {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.156437] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.connection = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.156603] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.connection_debug = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.156773] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.connection_parameters = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.156938] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.connection_recycle_time = 3600 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.157114] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.connection_trace = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.157283] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.db_inc_retry_interval = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.157446] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.db_max_retries = 20 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.157609] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.db_max_retry_interval = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.157771] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.db_retry_interval = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.157934] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.max_overflow = 50 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.158129] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.max_pool_size = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.158307] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.max_retries = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.158479] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.158639] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.158798] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.pool_timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.158961] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.retry_interval = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.159134] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.slave_connection = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.159301] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] api_database.sqlite_synchronous = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.159476] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] devices.enabled_mdev_types = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.159654] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.159825] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.159990] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ephemeral_storage_encryption.enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.160198] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.160379] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.api_servers = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.160545] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.160706] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.160871] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.161058] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.connect_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.161318] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.connect_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.161497] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.debug = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.161668] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.default_trusted_certificate_ids = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.161835] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.enable_certificate_validation = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.161999] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.enable_rbd_download = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.162175] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.endpoint_override = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.162347] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.162511] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.162669] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.max_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.162825] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.min_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.162989] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.num_retries = 3 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.163175] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.rbd_ceph_conf = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.163343] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.rbd_connect_timeout = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.163513] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.rbd_pool = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.163680] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.rbd_user = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.163840] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.region_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.164009] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.retriable_status_codes = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.164176] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.service_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.164346] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.service_type = image {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.164510] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.164669] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.status_code_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.164827] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.status_code_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.164987] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.165189] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.165360] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.verify_glance_signatures = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.165522] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] glance.version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.165689] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] guestfs.debug = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.165859] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] mks.enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.166227] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.166424] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] image_cache.manager_interval = 2400 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.166597] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] image_cache.precache_concurrency = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.166765] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] image_cache.remove_unused_base_images = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.166936] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.167119] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.167300] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] image_cache.subdirectory_name = _base {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.167477] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.api_max_retries = 60 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.167644] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.api_retry_interval = 2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.167805] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.auth_section = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.167968] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.auth_type = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.168169] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.168344] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.168511] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.168677] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.conductor_group = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.168840] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.connect_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.169012] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.connect_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.169180] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.endpoint_override = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.169346] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.169508] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.169668] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.max_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.169827] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.min_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.169993] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.peer_list = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.170200] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.region_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.170368] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.retriable_status_codes = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.170534] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.serial_console_state_timeout = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.170696] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.service_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.170868] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.service_type = baremetal {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.171053] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.shard = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.171314] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.171502] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.status_code_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.171666] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.status_code_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.171830] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.172019] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.172192] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ironic.version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.172374] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.172548] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] key_manager.fixed_key = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.172733] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.172895] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.barbican_api_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174100] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.barbican_endpoint = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174100] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.barbican_endpoint_type = public {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174100] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.barbican_region_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174100] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174100] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174100] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174100] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174338] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174448] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.number_of_retries = 60 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174615] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.retry_delay = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174784] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.send_service_user_token = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.174949] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.175126] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176022] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.verify_ssl = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176022] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican.verify_ssl_path = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176022] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican_service_user.auth_section = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176022] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican_service_user.auth_type = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176022] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican_service_user.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176252] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican_service_user.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176279] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican_service_user.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176433] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican_service_user.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176590] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican_service_user.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176757] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican_service_user.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.176918] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] barbican_service_user.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.177103] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.approle_role_id = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181034] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.approle_secret_id = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181034] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.kv_mountpoint = secret {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181034] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.kv_path = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181034] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.kv_version = 2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181034] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.namespace = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181034] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.root_token_id = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181034] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.ssl_ca_crt_file = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181303] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.timeout = 60.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181303] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.use_ssl = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181303] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181303] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.auth_section = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181303] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.auth_type = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181303] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181303] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181472] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181472] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.connect_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181472] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.connect_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181472] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.endpoint_override = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181472] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181472] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181472] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.max_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181643] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.min_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181643] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.region_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181643] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.retriable_status_codes = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181643] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.service_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181643] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.service_type = identity {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181643] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181793] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.status_code_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.181979] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.status_code_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.182104] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.182285] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.182443] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] keystone.version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.182641] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.connection_uri = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.182799] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.cpu_mode = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.182961] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.183151] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.cpu_models = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.183320] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.cpu_power_governor_high = performance {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.183484] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.183643] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.cpu_power_management = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.184049] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.184049] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.device_detach_attempts = 8 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.184212] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.device_detach_timeout = 20 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.184302] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.disk_cachemodes = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.184464] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.disk_prefix = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.184645] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.enabled_perf_events = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.184810] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.file_backed_memory = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.184976] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.gid_maps = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.185149] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.hw_disk_discard = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.185311] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.hw_machine_type = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.185479] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.images_rbd_ceph_conf = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.185644] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.185808] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.185979] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.images_rbd_glance_store_name = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.186169] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.images_rbd_pool = rbd {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.186348] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.images_type = default {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.186510] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.images_volume_group = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.186674] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.inject_key = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.186838] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.inject_partition = -2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.187007] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.inject_password = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.187178] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.iscsi_iface = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.187347] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.iser_use_multipath = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.187515] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.187679] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.187843] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_downtime = 500 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.188025] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.188213] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.188381] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_inbound_addr = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.188546] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.188709] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.188869] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_scheme = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.189055] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_timeout_action = abort {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.189233] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_tunnelled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.189397] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_uri = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.189564] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.live_migration_with_native_tls = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.189733] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.max_queues = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.189904] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.190173] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.190352] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.nfs_mount_options = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.190638] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.190813] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.190978] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.191178] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.191354] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.191592] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.num_pcie_ports = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.191787] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.191958] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.pmem_namespaces = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.192135] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.quobyte_client_cfg = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.192422] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.192596] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.192762] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.192924] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.193096] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.rbd_secret_uuid = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.193260] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.rbd_user = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.193422] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.193591] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.193748] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.rescue_image_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.193905] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.rescue_kernel_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.194072] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.rescue_ramdisk_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.194247] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.194406] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.rx_queue_size = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.194570] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.smbfs_mount_options = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.194840] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.195015] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.snapshot_compression = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.195182] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.snapshot_image_format = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.195404] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.195573] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.sparse_logical_volumes = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.195738] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.swtpm_enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.195905] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.swtpm_group = tss {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.196086] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.swtpm_user = tss {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.196263] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.sysinfo_serial = unique {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.196423] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.tb_cache_size = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.196583] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.tx_queue_size = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.196749] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.uid_maps = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.196913] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.use_virtio_for_bridges = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.197097] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.virt_type = kvm {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.197274] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.volume_clear = zero {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.197439] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.volume_clear_size = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.197604] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.volume_use_multipath = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.197763] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.vzstorage_cache_path = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.197931] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.198141] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.198325] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.198497] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.198768] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.198943] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.vzstorage_mount_user = stack {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.199124] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.199304] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.auth_section = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.199478] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.auth_type = password {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.199639] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.199797] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.199962] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.200169] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.connect_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.200344] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.connect_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.200517] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.default_floating_pool = public {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.200680] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.endpoint_override = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.200846] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.extension_sync_interval = 600 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.201026] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.http_retries = 3 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.201215] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.201378] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.201538] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.max_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.201791] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.201964] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.min_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.202153] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.ovs_bridge = br-int {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.202325] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.physnets = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.202497] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.region_name = RegionOne {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.202661] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.retriable_status_codes = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.202833] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.service_metadata_proxy = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.202994] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.service_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.203180] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.service_type = network {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.203348] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.203507] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.status_code_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.203668] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.status_code_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.203829] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.204014] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.204185] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] neutron.version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.204356] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] notifications.bdms_in_notifications = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.204533] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] notifications.default_level = INFO {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.204706] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] notifications.notification_format = unversioned {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.204869] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] notifications.notify_on_state_change = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.205055] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.205235] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] pci.alias = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.205404] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] pci.device_spec = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.205568] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] pci.report_in_placement = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.205740] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.auth_section = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.205911] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.auth_type = password {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.206088] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.206257] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.206415] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.206576] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.206734] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.connect_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.206894] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.connect_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.207060] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.default_domain_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.207221] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.default_domain_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.207378] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.domain_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.207533] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.domain_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.207686] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.endpoint_override = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.207846] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.208013] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.208203] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.max_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.208366] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.min_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.208535] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.password = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.208692] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.project_domain_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.208858] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.project_domain_name = Default {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.209034] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.project_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.209212] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.project_name = service {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.209383] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.region_name = RegionOne {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.209551] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.retriable_status_codes = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.209710] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.service_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.209876] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.service_type = placement {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.210066] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.210248] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.status_code_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.210415] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.status_code_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.210580] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.system_scope = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.210739] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.210897] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.trust_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.211082] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.user_domain_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.211268] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.user_domain_name = Default {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.211434] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.user_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.211606] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.username = nova {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.211865] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.212051] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] placement.version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.212238] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.cores = 20 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.212408] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.count_usage_from_placement = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.212582] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.212757] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.injected_file_content_bytes = 10240 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.212924] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.injected_file_path_length = 255 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.213103] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.injected_files = 5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.213276] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.instances = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.213442] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.key_pairs = 100 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.213610] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.metadata_items = 128 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.213774] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.ram = 51200 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.213936] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.recheck_quota = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.214119] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.server_group_members = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.214292] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] quota.server_groups = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.214465] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.214644] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.214819] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler.image_metadata_prefilter = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.214984] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.215163] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler.max_attempts = 3 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.215329] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler.max_placement_results = 1000 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.215494] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.215655] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.215818] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.215997] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] scheduler.workers = 2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.216185] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.216359] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.216538] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.216706] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.216870] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.217044] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.217214] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.217400] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.217571] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.host_subset_size = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.217737] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.217897] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.218090] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.218279] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.isolated_hosts = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.218451] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.isolated_images = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.218617] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.218788] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.218956] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.219136] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.pci_in_placement = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.219304] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.219465] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.219634] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.219790] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.219952] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.220154] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.220329] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.track_instance_changes = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.220507] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.220679] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] metrics.required = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.220842] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] metrics.weight_multiplier = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.221017] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.221212] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] metrics.weight_setting = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.221524] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.221700] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] serial_console.enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.221952] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] serial_console.port_range = 10000:20000 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.222157] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.222332] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.222500] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] serial_console.serialproxy_port = 6083 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.222667] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_user.auth_section = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.222840] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_user.auth_type = password {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.223009] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_user.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.223176] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_user.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.223342] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_user.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.223503] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_user.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.223661] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_user.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.223830] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_user.send_service_user_token = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.223993] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_user.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.224164] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] service_user.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.224335] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.agent_enabled = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.224497] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.224810] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.225007] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.225187] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.html5proxy_port = 6082 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.225353] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.image_compression = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.225512] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.jpeg_compression = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.225672] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.playback_compression = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.225834] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.require_secure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.226016] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.server_listen = 127.0.0.1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.226184] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.226345] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.streaming_mode = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.226500] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] spice.zlib_compression = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.226662] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] upgrade_levels.baseapi = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.226829] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] upgrade_levels.compute = auto {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.226986] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] upgrade_levels.conductor = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.227156] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] upgrade_levels.scheduler = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.227321] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.227481] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.227637] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vendordata_dynamic_auth.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.227792] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vendordata_dynamic_auth.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.227951] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.228146] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vendordata_dynamic_auth.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.228315] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.228481] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.228640] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vendordata_dynamic_auth.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.228812] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.api_retry_count = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.228971] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.ca_file = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.229163] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.229323] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.cluster_name = testcl1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.229485] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.connection_pool_size = 10 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.229641] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.console_delay_seconds = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.229806] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.datastore_regex = ^datastore.* {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.230023] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.230221] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.host_password = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.230394] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.host_port = 443 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.230561] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.host_username = administrator@vsphere.local {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.230726] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.insecure = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.230887] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.integration_bridge = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.231080] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.maximum_objects = 100 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.231260] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.pbm_default_policy = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.231429] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.pbm_enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.231589] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.pbm_wsdl_location = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.231760] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.231987] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.serial_port_proxy_uri = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.232184] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.serial_port_service_uri = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.232361] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.task_poll_interval = 0.5 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.232539] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.use_linked_clone = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.232711] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.vnc_keymap = en-us {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.232877] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.vnc_port = 5900 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.233052] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vmware.vnc_port_total = 10000 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.233244] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vnc.auth_schemes = ['none'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.233421] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vnc.enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.233710] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.233896] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.234078] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vnc.novncproxy_port = 6080 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.234261] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vnc.server_listen = 127.0.0.1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.234434] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.234595] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vnc.vencrypt_ca_certs = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.234754] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vnc.vencrypt_client_cert = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.234912] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vnc.vencrypt_client_key = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.235102] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.235270] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.disable_deep_image_inspection = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.235479] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.235672] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.235839] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.236008] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.disable_rootwrap = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.236180] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.enable_numa_live_migration = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.236347] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.236510] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.236670] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.236831] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.libvirt_disable_apic = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.236992] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.237169] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.237332] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.237495] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.237658] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.237819] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.237982] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.238183] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.238349] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.238515] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.238699] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.238867] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.client_socket_timeout = 900 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.239042] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.default_pool_size = 1000 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.239212] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.keep_alive = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.239375] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.max_header_line = 16384 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.239537] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.239696] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.ssl_ca_file = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.239853] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.ssl_cert_file = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.240021] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.ssl_key_file = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.240209] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.tcp_keepidle = 600 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.240392] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.240556] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] zvm.ca_file = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.240718] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] zvm.cloud_connector_url = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.240995] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.241203] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] zvm.reachable_timeout = 300 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.241393] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_policy.enforce_new_defaults = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.241775] env[61857]: WARNING oslo_config.cfg [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 480.241962] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_policy.enforce_scope = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.242228] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_policy.policy_default_rule = default {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.242427] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.242606] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_policy.policy_file = policy.yaml {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.242786] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.242951] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.243127] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.243294] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.243459] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.243628] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.243805] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.243983] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.connection_string = messaging:// {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.244168] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.enabled = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.244345] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.es_doc_type = notification {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.244511] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.es_scroll_size = 10000 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.244682] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.es_scroll_time = 2m {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.244845] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.filter_error_trace = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.245025] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.hmac_keys = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.245203] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.sentinel_service_name = mymaster {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.245371] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.socket_timeout = 0.1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.245536] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.trace_requests = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.245696] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler.trace_sqlalchemy = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.245871] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler_jaeger.process_tags = {} {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.246044] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler_jaeger.service_name_prefix = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.246215] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] profiler_otlp.service_name_prefix = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.246383] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] remote_debug.host = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.246543] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] remote_debug.port = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.246724] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.246890] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.247067] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.247238] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.247402] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.247562] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.247725] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.247887] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.248081] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.248274] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.248438] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.248611] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.248783] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.248955] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.249142] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.249316] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.249484] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.249659] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.249823] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.249986] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.250195] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.250371] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.250538] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.250709] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.250873] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.251046] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.251218] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.251384] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.251554] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.251721] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.ssl = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.251893] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.252075] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.252330] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.252514] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.252687] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.252852] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.253052] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.253229] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_notifications.retry = -1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.253416] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.253593] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.253766] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.auth_section = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.253933] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.auth_type = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.254109] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.cafile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.254276] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.certfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.254442] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.collect_timing = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.254603] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.connect_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.254764] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.connect_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.254925] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.endpoint_id = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.255096] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.endpoint_override = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.255266] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.insecure = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.255424] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.keyfile = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.255580] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.max_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.255737] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.min_version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.255895] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.region_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.256070] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.retriable_status_codes = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.256236] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.service_name = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.256396] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.service_type = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.256559] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.split_loggers = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.256718] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.status_code_retries = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.256880] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.status_code_retry_delay = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.257050] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.timeout = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.257217] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.valid_interfaces = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.257377] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_limit.version = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.257547] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_reports.file_event_handler = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.257708] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.257866] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] oslo_reports.log_dir = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.258063] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.258244] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.258406] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.258574] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.258739] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.258899] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.259088] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.259250] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_ovs_privileged.group = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.259410] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.259576] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.259741] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.259899] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] vif_plug_ovs_privileged.user = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.260097] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.260289] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.260464] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.260637] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.260805] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.260972] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.261152] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.261319] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.261496] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.261666] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_ovs.isolate_vif = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.261832] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.261998] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.262186] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.262442] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.262616] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_vif_ovs.per_port_bridge = False {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.262786] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_brick.lock_path = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.262953] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.263132] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.263307] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] privsep_osbrick.capabilities = [21] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.263469] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] privsep_osbrick.group = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.263629] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] privsep_osbrick.helper_command = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.263795] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.263958] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.264131] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] privsep_osbrick.user = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.264306] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.264465] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] nova_sys_admin.group = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.264623] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] nova_sys_admin.helper_command = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.264785] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.264948] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.265119] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] nova_sys_admin.user = None {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 480.265254] env[61857]: DEBUG oslo_service.service [None req-5fbfd3d8-120f-4665-8f13-6b75223f6f25 None None] ******************************************************************************** {{(pid=61857) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 480.265739] env[61857]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 480.768649] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Getting list of instances from cluster (obj){ [ 480.768649] env[61857]: value = "domain-c8" [ 480.768649] env[61857]: _type = "ClusterComputeResource" [ 480.768649] env[61857]: } {{(pid=61857) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 480.769814] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a81eaab-3caf-47a7-925a-646882de4ac5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.780169] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Got total of 0 instances {{(pid=61857) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 480.780169] env[61857]: WARNING nova.virt.vmwareapi.driver [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 480.780330] env[61857]: INFO nova.virt.node [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Generated node identity 2d5860d0-8574-4e55-9ced-20e66f0314c2 [ 480.780492] env[61857]: INFO nova.virt.node [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Wrote node identity 2d5860d0-8574-4e55-9ced-20e66f0314c2 to /opt/stack/data/n-cpu-1/compute_id [ 481.283210] env[61857]: WARNING nova.compute.manager [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Compute nodes ['2d5860d0-8574-4e55-9ced-20e66f0314c2'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 482.290384] env[61857]: INFO nova.compute.manager [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 483.298175] env[61857]: WARNING nova.compute.manager [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 483.298537] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 483.298696] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 483.298857] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 483.299043] env[61857]: DEBUG nova.compute.resource_tracker [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 483.299982] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f5b43d-0546-4719-bf1b-8a2845319ef4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.308328] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39f3311-7d28-4861-b444-3ef97ff6bbd6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.321545] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c4eaec-8972-429e-8e58-3f887f700ac7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.327657] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1051014-4fd7-40c9-b8fb-3e9aeb13dcb2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.356748] env[61857]: DEBUG nova.compute.resource_tracker [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181649MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 483.356912] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 483.357080] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 483.860023] env[61857]: WARNING nova.compute.resource_tracker [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] No compute node record for cpu-1:2d5860d0-8574-4e55-9ced-20e66f0314c2: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 2d5860d0-8574-4e55-9ced-20e66f0314c2 could not be found. [ 484.363444] env[61857]: INFO nova.compute.resource_tracker [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 2d5860d0-8574-4e55-9ced-20e66f0314c2 [ 485.871422] env[61857]: DEBUG nova.compute.resource_tracker [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 485.871862] env[61857]: DEBUG nova.compute.resource_tracker [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 486.030840] env[61857]: INFO nova.scheduler.client.report [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] [req-79e327e5-7cae-42aa-8f3a-ffc136758bf7] Created resource provider record via placement API for resource provider with UUID 2d5860d0-8574-4e55-9ced-20e66f0314c2 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 486.048767] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcee90d0-4857-435f-bbc9-648a8ad7b480 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 486.056501] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1162c0f-f174-42b2-a582-08481079dac3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 486.086559] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79963d08-87de-4198-8c72-2389a5eaae40 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 486.093680] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b59907-e969-4763-9cec-8056331dd1ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 486.106389] env[61857]: DEBUG nova.compute.provider_tree [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 486.641954] env[61857]: DEBUG nova.scheduler.client.report [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Updated inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 486.642197] env[61857]: DEBUG nova.compute.provider_tree [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Updating resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 generation from 0 to 1 during operation: update_inventory {{(pid=61857) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 486.642339] env[61857]: DEBUG nova.compute.provider_tree [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 486.694872] env[61857]: DEBUG nova.compute.provider_tree [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Updating resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 generation from 1 to 2 during operation: update_traits {{(pid=61857) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 487.199600] env[61857]: DEBUG nova.compute.resource_tracker [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 487.199965] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.843s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 487.199965] env[61857]: DEBUG nova.service [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Creating RPC server for service compute {{(pid=61857) start /opt/stack/nova/nova/service.py:186}} [ 487.213928] env[61857]: DEBUG nova.service [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] Join ServiceGroup membership for this service compute {{(pid=61857) start /opt/stack/nova/nova/service.py:203}} [ 487.214143] env[61857]: DEBUG nova.servicegroup.drivers.db [None req-10a92c1e-d340-4288-88dd-de303b6adb6a None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61857) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 491.216455] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 491.719576] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Getting list of instances from cluster (obj){ [ 491.719576] env[61857]: value = "domain-c8" [ 491.719576] env[61857]: _type = "ClusterComputeResource" [ 491.719576] env[61857]: } {{(pid=61857) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 491.720863] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e62673-e8a6-4a6d-83ee-66150dc5e953 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.729052] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Got total of 0 instances {{(pid=61857) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 491.729353] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 491.729695] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Getting list of instances from cluster (obj){ [ 491.729695] env[61857]: value = "domain-c8" [ 491.729695] env[61857]: _type = "ClusterComputeResource" [ 491.729695] env[61857]: } {{(pid=61857) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 491.730566] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38bbbc66-5f33-46ca-9bdc-cd62172ec97b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.737480] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Got total of 0 instances {{(pid=61857) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 519.967185] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "bb418d62-d47c-42c5-908e-26f1eb034e46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.967185] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "bb418d62-d47c-42c5-908e-26f1eb034e46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.474631] env[61857]: DEBUG nova.compute.manager [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 521.023772] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.024074] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.025634] env[61857]: INFO nova.compute.claims [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.083174] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739145f6-30f8-45db-bf4c-14f62ab25fdb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.092353] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d68163-a1e3-4667-9443-f536c9a64ff1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.129618] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1369d24c-cd69-4c23-84ef-b5ff85289491 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.140309] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45babd66-0eae-4d31-b17d-971298f5cd92 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.154965] env[61857]: DEBUG nova.compute.provider_tree [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.658560] env[61857]: DEBUG nova.scheduler.client.report [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 523.164296] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.140s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.165136] env[61857]: DEBUG nova.compute.manager [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 523.673684] env[61857]: DEBUG nova.compute.utils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.676121] env[61857]: DEBUG nova.compute.manager [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Not allocating networking since 'none' was specified. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 524.179114] env[61857]: DEBUG nova.compute.manager [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 525.138076] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquiring lock "61faaebe-374a-40ba-aefb-b199ea4dea93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.138347] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "61faaebe-374a-40ba-aefb-b199ea4dea93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.194846] env[61857]: DEBUG nova.compute.manager [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 525.312050] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 525.312050] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 525.312050] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 525.312820] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 525.313788] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 525.315392] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 525.317865] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 525.317865] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 525.317865] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 525.317865] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 525.317865] env[61857]: DEBUG nova.virt.hardware [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 525.320042] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada4c2cf-b6cb-431e-8a49-77000904fd79 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.334582] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a5a754-9776-4f73-8723-1117a98b55d7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.352818] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ab4c24-c1a9-43b7-8d7d-9a10efcf1624 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.372882] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 525.383444] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 525.383781] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d7f2524c-cde6-4ffd-910e-e56831140b0e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.399138] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Created folder: OpenStack in parent group-v4. [ 525.399430] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Creating folder: Project (3b1fb545dd1e4788af2cfeb2ecb54bf3). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 525.399833] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d7694aa-a707-49cc-a835-1e414ea21654 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.409625] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Created folder: Project (3b1fb545dd1e4788af2cfeb2ecb54bf3) in parent group-v214027. [ 525.409826] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Creating folder: Instances. Parent ref: group-v214028. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 525.410090] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08517eb2-1337-485f-b7ac-a66817fbfd16 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.422164] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Created folder: Instances in parent group-v214028. [ 525.422164] env[61857]: DEBUG oslo.service.loopingcall [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.423528] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 525.426747] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cbf34fc2-114c-474c-a1a0-1fc97c8a38e7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.441833] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquiring lock "90b0c4ad-902e-425d-b66b-51918e6476d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.441833] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "90b0c4ad-902e-425d-b66b-51918e6476d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.449823] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 525.449823] env[61857]: value = "task-950655" [ 525.449823] env[61857]: _type = "Task" [ 525.449823] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.463220] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950655, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.641873] env[61857]: DEBUG nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.945575] env[61857]: DEBUG nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.969272] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950655, 'name': CreateVM_Task, 'duration_secs': 0.292464} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.969861] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 525.971189] env[61857]: DEBUG oslo_vmware.service [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7904d7e-6155-44c4-8860-99e15516554c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.979224] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.979224] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.979787] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 525.980211] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abb2619c-1259-4988-b8d6-5b43eae4c04a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.985447] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 525.985447] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524cc1c3-6144-84d1-4ee6-cddd0b45eb7d" [ 525.985447] env[61857]: _type = "Task" [ 525.985447] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.994041] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524cc1c3-6144-84d1-4ee6-cddd0b45eb7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.176671] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.177265] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.179980] env[61857]: INFO nova.compute.claims [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 526.363089] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Acquiring lock "b34ca3cc-d26d-4068-a27d-a6d66404dfcb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.363336] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Lock "b34ca3cc-d26d-4068-a27d-a6d66404dfcb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.483795] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.497691] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.497691] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 526.497691] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.497946] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.498590] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 526.498842] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e12c182-3346-4f3c-b74b-7e1811b3d66a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.518625] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 526.518625] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 526.518625] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd3da43-a58d-407d-a552-c2671e8c241d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.527966] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e073790-dd17-4b0a-88f4-024cdf623ec6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.533575] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 526.533575] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52277b54-67b2-d9e1-4c0a-6a41e31aacf3" [ 526.533575] env[61857]: _type = "Task" [ 526.533575] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.542413] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52277b54-67b2-d9e1-4c0a-6a41e31aacf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.709363] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Acquiring lock "6acac8fe-a2a7-426e-912a-0292d351887e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.709623] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Lock "6acac8fe-a2a7-426e-912a-0292d351887e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.822866] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6417baa7-a777-4788-9653-b5b0d0ba2837 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.833130] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d857704-3ddb-4c3f-bd4b-bb9e4d1e734d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.868497] env[61857]: DEBUG nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 526.871626] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c23723-ea56-4ae8-adbc-5093040e5d2a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.879139] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4c63f4-a088-4ff0-8f9e-f19aeda5326b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.896512] env[61857]: DEBUG nova.compute.provider_tree [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.050517] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Preparing fetch location {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 527.050781] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Creating directory with path [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 527.051099] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a1ec61f-82ea-4576-9c1e-f642a8937d09 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.073945] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Created directory with path [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 527.074181] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Fetch image to [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 527.074351] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Downloading image file data 2e0c8bb4-0827-486c-98ce-7994555ffc23 to [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk on the data store datastore2 {{(pid=61857) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 527.075200] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea92a3c-04eb-44cb-843e-e8b5156ae538 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.084622] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23cba9b-534b-47b9-8752-f7950be064b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.098557] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd1a634b-cbd5-445b-8098-2e078456fb6e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.136543] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c453c290-7799-4235-9a57-0004d8e68c7c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.143114] env[61857]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-29fca54c-d534-4bee-b5a0-38a14734dfb6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.165741] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Downloading image file data 2e0c8bb4-0827-486c-98ce-7994555ffc23 to the data store datastore2 {{(pid=61857) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 527.217175] env[61857]: DEBUG nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 527.271203] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.271445] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.380255] env[61857]: DEBUG oslo_vmware.rw_handles [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61857) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 527.455980] env[61857]: DEBUG nova.scheduler.client.report [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.461604] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.752290] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.773447] env[61857]: DEBUG nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 527.965162] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.788s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.966278] env[61857]: DEBUG nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 527.974403] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.491s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.976596] env[61857]: INFO nova.compute.claims [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 528.070186] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Acquiring lock "4c7dab9e-4527-46b6-a1ec-6e6935768b75" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.070431] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Lock "4c7dab9e-4527-46b6-a1ec-6e6935768b75" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.088164] env[61857]: DEBUG oslo_vmware.rw_handles [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Completed reading data from the image iterator. {{(pid=61857) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 528.088164] env[61857]: DEBUG oslo_vmware.rw_handles [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 528.221281] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Downloaded image file data 2e0c8bb4-0827-486c-98ce-7994555ffc23 to vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk on the data store datastore2 {{(pid=61857) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 528.223094] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Caching image {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 528.223521] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Copying Virtual Disk [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk to [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 528.224621] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f71e4f8-10e1-44bd-a71f-2aa5dd7d2f21 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.234662] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 528.234662] env[61857]: value = "task-950656" [ 528.234662] env[61857]: _type = "Task" [ 528.234662] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.245958] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950656, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.312405] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.486799] env[61857]: DEBUG nova.compute.utils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 528.487881] env[61857]: DEBUG nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 528.488363] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 528.573089] env[61857]: DEBUG nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.757246] env[61857]: DEBUG nova.policy [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3317265525e1499b87e847f2ad30ed60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '883ec0e28c9040dd89202cd07932ea7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.770928] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950656, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.008022] env[61857]: DEBUG nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 529.108351] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.173758] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668a7cec-7a09-48ed-84bc-fde95e59b26d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.184257] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f43a34-3a84-4ce3-be4a-7a33b9dcabd8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.225370] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd822c52-88f5-40ae-b799-cfbce44221d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.236625] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a645625-e05a-45ae-b9a2-37f38977d903 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.252564] env[61857]: DEBUG nova.compute.provider_tree [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.274169] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950656, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.734007} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.274446] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Copied Virtual Disk [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk to [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 529.275973] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleting the datastore file [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 529.276278] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37e0f242-e944-4654-8ae3-71841ed5b9ed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.283796] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 529.283796] env[61857]: value = "task-950657" [ 529.283796] env[61857]: _type = "Task" [ 529.283796] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.292089] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950657, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.756218] env[61857]: DEBUG nova.scheduler.client.report [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.804229] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950657, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023053} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.804481] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 529.804685] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Moving file from [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec/2e0c8bb4-0827-486c-98ce-7994555ffc23 to [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23. {{(pid=61857) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 529.804933] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-a9fc1819-e8b7-4983-bb4c-7e49f82faf4c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.820823] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 529.820823] env[61857]: value = "task-950658" [ 529.820823] env[61857]: _type = "Task" [ 529.820823] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.835576] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950658, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.836859] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Successfully created port: ea2b2110-9959-44fa-a150-e11ae8904d4c {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 530.023530] env[61857]: DEBUG nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 530.055415] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 530.055657] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 530.055807] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 530.055984] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 530.056440] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 530.056607] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 530.056834] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 530.056990] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 530.057168] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 530.057325] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 530.057486] env[61857]: DEBUG nova.virt.hardware [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 530.058378] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64caa560-33e3-4a14-99a8-3741ddabf276 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.069722] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b031216-a7c3-427c-8666-f55fc9f60f50 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.265105] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.290s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.265105] env[61857]: DEBUG nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 530.268296] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.806s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.270284] env[61857]: INFO nova.compute.claims [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 530.339093] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950658, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024984} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.339093] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] File moved {{(pid=61857) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 530.339093] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Cleaning up location [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 530.339093] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleting the datastore file [datastore2] vmware_temp/9bccbab7-b840-4018-90c5-f21933cef4ec {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 530.339093] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7545893b-0458-4423-83c4-5fbf76afeda5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.345924] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 530.345924] env[61857]: value = "task-950659" [ 530.345924] env[61857]: _type = "Task" [ 530.345924] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.357245] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950659, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.780215] env[61857]: DEBUG nova.compute.utils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.781914] env[61857]: DEBUG nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 530.782114] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 530.859663] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950659, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.043373} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.860210] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 530.862772] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bb77594-2ddd-4666-a9fa-2658ebd63c87 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.868145] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 530.868145] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5290908c-d83f-d4f7-ddf3-fd369a075ac8" [ 530.868145] env[61857]: _type = "Task" [ 530.868145] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.876226] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5290908c-d83f-d4f7-ddf3-fd369a075ac8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.094620] env[61857]: DEBUG nova.policy [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8722a9ebebd14eacac59c7812251de79', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd59a31d1f9740fb8f653b4406729a3a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 531.286138] env[61857]: DEBUG nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 531.386481] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5290908c-d83f-d4f7-ddf3-fd369a075ac8, 'name': SearchDatastore_Task, 'duration_secs': 0.009061} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.388473] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.388736] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 531.389619] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b90e7679-5cdb-4b0f-9fc8-24144265f28d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.395800] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 531.395800] env[61857]: value = "task-950660" [ 531.395800] env[61857]: _type = "Task" [ 531.395800] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.407230] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950660, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.451517] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a279a72-2c52-4348-a425-765f3642bd61 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.462259] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa92812-6cb9-4e9d-8243-f487f8cd77f1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.498998] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a324c4b-bea0-4282-a15a-316470b9b0b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.508454] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45f1555-cd8e-4140-be2f-6c4afe952f10 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.536560] env[61857]: DEBUG nova.compute.provider_tree [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.908413] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950660, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468045} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.911028] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 531.911028] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 531.911028] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47ec0789-a019-41f6-9b88-3f276cdcfbeb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.919325] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 531.919325] env[61857]: value = "task-950661" [ 531.919325] env[61857]: _type = "Task" [ 531.919325] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.930071] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950661, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.038169] env[61857]: DEBUG nova.scheduler.client.report [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 532.297352] env[61857]: DEBUG nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 532.325339] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 532.325801] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 532.325801] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.326270] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 532.326270] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.326270] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 532.326430] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 532.326541] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 532.327114] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 532.327478] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 532.328194] env[61857]: DEBUG nova.virt.hardware [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 532.330413] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f688bcc-9b45-486c-aeb1-4ecb6f5e41b9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.337670] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4cbf44c-d288-422b-b2b6-4ab963034829 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.428912] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950661, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071082} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.429205] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 532.430175] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ab5328-6094-4310-bf2e-49236f519211 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.457211] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 532.457543] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8447929c-4110-4741-886c-9cd3e967a9f8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.481069] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 532.481069] env[61857]: value = "task-950662" [ 532.481069] env[61857]: _type = "Task" [ 532.481069] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.491868] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950662, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.546464] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.547169] env[61857]: DEBUG nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 532.549859] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.799s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.551284] env[61857]: INFO nova.compute.claims [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 532.885262] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Successfully created port: 4b0da35d-33e2-494e-a2da-473d4e03a462 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.991566] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950662, 'name': ReconfigVM_Task, 'duration_secs': 0.302287} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.994119] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Reconfigured VM instance instance-00000001 to attach disk [datastore2] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 532.994842] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1fd467ba-e269-4ae9-8e07-7362cd01298a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.002996] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 533.002996] env[61857]: value = "task-950663" [ 533.002996] env[61857]: _type = "Task" [ 533.002996] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.013210] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950663, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.059817] env[61857]: DEBUG nova.compute.utils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 533.063665] env[61857]: DEBUG nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 533.063925] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 533.375378] env[61857]: DEBUG nova.policy [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5edc51f95eac44158c13a7d1ed8ff2d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac067183bb7d4f358f1d34cb66debe33', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 533.513312] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950663, 'name': Rename_Task, 'duration_secs': 0.172033} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.513600] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 533.513848] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-50eba9e7-00ef-4682-bd16-355a44d17e9b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.520366] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 533.520366] env[61857]: value = "task-950664" [ 533.520366] env[61857]: _type = "Task" [ 533.520366] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.528474] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.564235] env[61857]: DEBUG nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 533.723332] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3c22cd-10d8-4434-a7e6-6733ec1aa65f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.731508] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3f8772-5453-4a62-843f-ad6014b5b13c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.766620] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f91ef8b-fa89-4ab1-9815-d7031a55b23d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.781385] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39bd000-7bf2-42e1-a116-9a49fa1c5a3c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.805263] env[61857]: DEBUG nova.compute.provider_tree [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.032180] env[61857]: DEBUG oslo_vmware.api [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950664, 'name': PowerOnVM_Task, 'duration_secs': 0.466722} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.032508] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 534.032821] env[61857]: INFO nova.compute.manager [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Took 8.84 seconds to spawn the instance on the hypervisor. [ 534.033164] env[61857]: DEBUG nova.compute.manager [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 534.033875] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e301fb5-5041-49cb-93a2-4afcad5abfbe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.308704] env[61857]: DEBUG nova.scheduler.client.report [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 534.557553] env[61857]: INFO nova.compute.manager [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Took 13.58 seconds to build instance. [ 534.575501] env[61857]: DEBUG nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 534.612102] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.612611] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.612877] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.613139] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.613704] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.613889] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.614127] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.614294] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.615317] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.615317] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.615317] env[61857]: DEBUG nova.virt.hardware [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.616581] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3368f7-128a-4fc7-8519-c73699e85773 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.625749] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850fba2b-2165-49e4-9792-ac05a22b7087 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.816434] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.264s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.816434] env[61857]: DEBUG nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.819425] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.507s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.822609] env[61857]: INFO nova.compute.claims [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 535.060672] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0f3712ed-f022-4b30-88eb-e46940e776af tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "bb418d62-d47c-42c5-908e-26f1eb034e46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.093s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.329495] env[61857]: DEBUG nova.compute.utils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 535.338358] env[61857]: DEBUG nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 535.339330] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 535.382620] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Successfully created port: 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 535.591695] env[61857]: DEBUG nova.policy [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b85ae2941ede400aba5feef20be01553', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '984afb02484a454f852c751e97001805', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 535.697441] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 535.698633] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 535.698998] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 535.699395] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Rebuilding the list of instances to heal {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 535.838889] env[61857]: DEBUG nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 535.984777] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b98e98a-462b-411b-b984-3351533129a4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.994007] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d28078-a2c7-4726-a1b8-109710a0056c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.026629] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c2e2d5-63e2-4d57-a26b-cddcc4ed63ff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.034874] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9816760f-8b50-4bcf-9b7a-c2c1ac2e5fa5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.049954] env[61857]: DEBUG nova.compute.provider_tree [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 536.204503] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 536.204503] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 536.204747] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 536.204747] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 536.204910] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 536.279898] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "refresh_cache-bb418d62-d47c-42c5-908e-26f1eb034e46" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.280124] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquired lock "refresh_cache-bb418d62-d47c-42c5-908e-26f1eb034e46" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.280278] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Forcefully refreshing network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 536.280463] env[61857]: DEBUG nova.objects.instance [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lazy-loading 'info_cache' on Instance uuid bb418d62-d47c-42c5-908e-26f1eb034e46 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 536.552913] env[61857]: DEBUG nova.scheduler.client.report [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 536.848092] env[61857]: DEBUG nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.879640] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.879867] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.880033] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.881932] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.881932] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.881932] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.881932] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.881932] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.882233] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.882233] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.882233] env[61857]: DEBUG nova.virt.hardware [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.885765] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c5fe74-1f76-4c2a-9efa-a6dbe1305756 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.894482] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b89568-7b16-4677-81f9-f590f1881d2d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.061764] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.242s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.062022] env[61857]: DEBUG nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 537.065337] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.957s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.067758] env[61857]: INFO nova.compute.claims [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.139992] env[61857]: ERROR nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ea2b2110-9959-44fa-a150-e11ae8904d4c, please check neutron logs for more information. [ 537.139992] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.139992] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.139992] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.139992] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.139992] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.139992] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.139992] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.139992] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.139992] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 537.139992] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.139992] env[61857]: ERROR nova.compute.manager raise self.value [ 537.139992] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.139992] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.139992] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.139992] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.140542] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.140542] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.140542] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ea2b2110-9959-44fa-a150-e11ae8904d4c, please check neutron logs for more information. [ 537.140542] env[61857]: ERROR nova.compute.manager [ 537.140542] env[61857]: Traceback (most recent call last): [ 537.140542] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.140542] env[61857]: listener.cb(fileno) [ 537.140542] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.140542] env[61857]: result = function(*args, **kwargs) [ 537.140542] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.140542] env[61857]: return func(*args, **kwargs) [ 537.140542] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.140542] env[61857]: raise e [ 537.140542] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.140542] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 537.140542] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.140542] env[61857]: created_port_ids = self._update_ports_for_instance( [ 537.140542] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.140542] env[61857]: with excutils.save_and_reraise_exception(): [ 537.140542] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.140542] env[61857]: self.force_reraise() [ 537.140542] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.140542] env[61857]: raise self.value [ 537.140542] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.140542] env[61857]: updated_port = self._update_port( [ 537.140542] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.140542] env[61857]: _ensure_no_port_binding_failure(port) [ 537.140542] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.140542] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.141320] env[61857]: nova.exception.PortBindingFailed: Binding failed for port ea2b2110-9959-44fa-a150-e11ae8904d4c, please check neutron logs for more information. [ 537.141320] env[61857]: Removing descriptor: 14 [ 537.143626] env[61857]: ERROR nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ea2b2110-9959-44fa-a150-e11ae8904d4c, please check neutron logs for more information. [ 537.143626] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Traceback (most recent call last): [ 537.143626] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 537.143626] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] yield resources [ 537.143626] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.143626] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] self.driver.spawn(context, instance, image_meta, [ 537.143626] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 537.143626] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.143626] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.143626] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] vm_ref = self.build_virtual_machine(instance, [ 537.143626] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] for vif in network_info: [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] return self._sync_wrapper(fn, *args, **kwargs) [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] self.wait() [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] self[:] = self._gt.wait() [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] return self._exit_event.wait() [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.143985] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] result = hub.switch() [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] return self.greenlet.switch() [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] result = function(*args, **kwargs) [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] return func(*args, **kwargs) [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] raise e [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] nwinfo = self.network_api.allocate_for_instance( [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] created_port_ids = self._update_ports_for_instance( [ 537.144322] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] with excutils.save_and_reraise_exception(): [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] self.force_reraise() [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] raise self.value [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] updated_port = self._update_port( [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] _ensure_no_port_binding_failure(port) [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] raise exception.PortBindingFailed(port_id=port['id']) [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] nova.exception.PortBindingFailed: Binding failed for port ea2b2110-9959-44fa-a150-e11ae8904d4c, please check neutron logs for more information. [ 537.144658] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] [ 537.147173] env[61857]: INFO nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Terminating instance [ 537.148756] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquiring lock "refresh_cache-61faaebe-374a-40ba-aefb-b199ea4dea93" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.148942] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquired lock "refresh_cache-61faaebe-374a-40ba-aefb-b199ea4dea93" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.148999] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 537.381694] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.572263] env[61857]: DEBUG nova.compute.utils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 537.577922] env[61857]: DEBUG nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 537.578437] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 537.732339] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.902522] env[61857]: DEBUG nova.policy [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac8387df3290404283263440672de653', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3bbcf7c2937040e1906e2273a07b671b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 538.083566] env[61857]: DEBUG nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 538.112256] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.241752] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Successfully created port: 27d6f88e-1abe-4552-b7a1-839802105483 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 538.280774] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ddeda5-8b8b-4dda-adbf-b062c9d13e6c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.290998] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abc6d2f-b9ee-4076-bb86-bd362eeb6dc8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.332790] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.337714] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67e9579-e147-4990-97bd-cb1850f629b2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.343310] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "67ff728e-37b1-4633-bfd2-75fe4b1c7d13" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.343379] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "67ff728e-37b1-4633-bfd2-75fe4b1c7d13" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.351318] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9442d930-d9b5-4af8-b86b-3978943f286e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.367984] env[61857]: DEBUG nova.compute.provider_tree [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.619534] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Releasing lock "refresh_cache-61faaebe-374a-40ba-aefb-b199ea4dea93" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.622771] env[61857]: DEBUG nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 538.622771] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 538.622771] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c160aaa-684e-4bc0-b354-114ea3a4d3a4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.634753] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60be24ec-b224-4001-a0eb-092761e5f53a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.660666] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 61faaebe-374a-40ba-aefb-b199ea4dea93 could not be found. [ 538.660890] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 538.661081] env[61857]: INFO nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Took 0.04 seconds to destroy the instance on the hypervisor. [ 538.661331] env[61857]: DEBUG oslo.service.loopingcall [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.661658] env[61857]: DEBUG nova.compute.manager [-] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.661759] env[61857]: DEBUG nova.network.neutron [-] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 538.714133] env[61857]: DEBUG nova.network.neutron [-] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.843966] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Releasing lock "refresh_cache-bb418d62-d47c-42c5-908e-26f1eb034e46" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.844729] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Updated the network info_cache for instance {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 538.845208] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 538.845623] env[61857]: DEBUG nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 538.849802] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 538.850480] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 538.850938] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 538.850938] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 538.851639] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 538.851639] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 538.851639] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 538.871139] env[61857]: DEBUG nova.scheduler.client.report [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.099150] env[61857]: DEBUG nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 539.131237] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 539.131237] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 539.131237] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 539.131515] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 539.132176] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 539.132586] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 539.132972] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 539.133310] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 539.133632] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 539.133934] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 539.134357] env[61857]: DEBUG nova.virt.hardware [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 539.135696] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10cae629-32e2-4040-8ee5-ed0c34596385 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.144988] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744b35f2-86df-439e-91a7-cdaef7f646d3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.218315] env[61857]: DEBUG nova.network.neutron [-] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.334431] env[61857]: ERROR nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4b0da35d-33e2-494e-a2da-473d4e03a462, please check neutron logs for more information. [ 539.334431] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 539.334431] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.334431] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 539.334431] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.334431] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 539.334431] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.334431] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 539.334431] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.334431] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 539.334431] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.334431] env[61857]: ERROR nova.compute.manager raise self.value [ 539.334431] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.334431] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 539.334431] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.334431] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 539.335172] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.335172] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 539.335172] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4b0da35d-33e2-494e-a2da-473d4e03a462, please check neutron logs for more information. [ 539.335172] env[61857]: ERROR nova.compute.manager [ 539.335172] env[61857]: Traceback (most recent call last): [ 539.335172] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 539.335172] env[61857]: listener.cb(fileno) [ 539.335172] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.335172] env[61857]: result = function(*args, **kwargs) [ 539.335172] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 539.335172] env[61857]: return func(*args, **kwargs) [ 539.335172] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.335172] env[61857]: raise e [ 539.335172] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.335172] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 539.335172] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.335172] env[61857]: created_port_ids = self._update_ports_for_instance( [ 539.335172] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.335172] env[61857]: with excutils.save_and_reraise_exception(): [ 539.335172] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.335172] env[61857]: self.force_reraise() [ 539.335172] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.335172] env[61857]: raise self.value [ 539.335172] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.335172] env[61857]: updated_port = self._update_port( [ 539.335172] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.335172] env[61857]: _ensure_no_port_binding_failure(port) [ 539.335172] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.335172] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 539.337362] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 4b0da35d-33e2-494e-a2da-473d4e03a462, please check neutron logs for more information. [ 539.337362] env[61857]: Removing descriptor: 16 [ 539.337362] env[61857]: ERROR nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4b0da35d-33e2-494e-a2da-473d4e03a462, please check neutron logs for more information. [ 539.337362] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Traceback (most recent call last): [ 539.337362] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 539.337362] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] yield resources [ 539.337362] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 539.337362] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] self.driver.spawn(context, instance, image_meta, [ 539.337362] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 539.337362] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 539.337362] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 539.337362] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] vm_ref = self.build_virtual_machine(instance, [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] for vif in network_info: [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] return self._sync_wrapper(fn, *args, **kwargs) [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] self.wait() [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] self[:] = self._gt.wait() [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] return self._exit_event.wait() [ 539.337890] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] result = hub.switch() [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] return self.greenlet.switch() [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] result = function(*args, **kwargs) [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] return func(*args, **kwargs) [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] raise e [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] nwinfo = self.network_api.allocate_for_instance( [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.338226] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] created_port_ids = self._update_ports_for_instance( [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] with excutils.save_and_reraise_exception(): [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] self.force_reraise() [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] raise self.value [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] updated_port = self._update_port( [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] _ensure_no_port_binding_failure(port) [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.338534] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] raise exception.PortBindingFailed(port_id=port['id']) [ 539.338822] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] nova.exception.PortBindingFailed: Binding failed for port 4b0da35d-33e2-494e-a2da-473d4e03a462, please check neutron logs for more information. [ 539.338822] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] [ 539.338822] env[61857]: INFO nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Terminating instance [ 539.338822] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquiring lock "refresh_cache-90b0c4ad-902e-425d-b66b-51918e6476d9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.338822] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquired lock "refresh_cache-90b0c4ad-902e-425d-b66b-51918e6476d9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.338822] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 539.354373] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.375184] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.375716] env[61857]: DEBUG nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 539.381199] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.025s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.381199] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.381199] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 539.381199] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b4be7d-7d89-4bcd-afdb-2808e8e65344 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.385036] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.385141] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.387116] env[61857]: INFO nova.compute.claims [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 539.396539] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4fe5c2-459c-4b77-96b9-f6eca11a3461 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.416327] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa8dd1e-252e-4b0d-bc11-ae9bfb071836 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.425226] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3e81d7-d089-41e2-a14f-1d2cb0991fb0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.460111] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181647MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 539.460111] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.461271] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Successfully created port: bba41198-8b1a-4895-ae13-fd458fd4eb5d {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 539.720971] env[61857]: INFO nova.compute.manager [-] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Took 1.06 seconds to deallocate network for instance. [ 539.730259] env[61857]: DEBUG nova.compute.claims [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.730931] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.891861] env[61857]: DEBUG nova.compute.utils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 539.897780] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.902231] env[61857]: DEBUG nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 539.902231] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 540.036264] env[61857]: DEBUG nova.compute.manager [req-637b63a6-fc58-458f-9663-fd1f65913f31 req-269346f0-ed34-4b46-9205-4a7b3ede932e service nova] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Received event network-changed-ea2b2110-9959-44fa-a150-e11ae8904d4c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 540.036445] env[61857]: DEBUG nova.compute.manager [req-637b63a6-fc58-458f-9663-fd1f65913f31 req-269346f0-ed34-4b46-9205-4a7b3ede932e service nova] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Refreshing instance network info cache due to event network-changed-ea2b2110-9959-44fa-a150-e11ae8904d4c. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 540.036660] env[61857]: DEBUG oslo_concurrency.lockutils [req-637b63a6-fc58-458f-9663-fd1f65913f31 req-269346f0-ed34-4b46-9205-4a7b3ede932e service nova] Acquiring lock "refresh_cache-61faaebe-374a-40ba-aefb-b199ea4dea93" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.036801] env[61857]: DEBUG oslo_concurrency.lockutils [req-637b63a6-fc58-458f-9663-fd1f65913f31 req-269346f0-ed34-4b46-9205-4a7b3ede932e service nova] Acquired lock "refresh_cache-61faaebe-374a-40ba-aefb-b199ea4dea93" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.036984] env[61857]: DEBUG nova.network.neutron [req-637b63a6-fc58-458f-9663-fd1f65913f31 req-269346f0-ed34-4b46-9205-4a7b3ede932e service nova] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Refreshing network info cache for port ea2b2110-9959-44fa-a150-e11ae8904d4c {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 540.098374] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Acquiring lock "ad56d32b-c78d-45db-872a-1c4ea91ab909" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.098600] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Lock "ad56d32b-c78d-45db-872a-1c4ea91ab909" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.164335] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.174286] env[61857]: DEBUG nova.policy [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c0dd112eec2c4e538f61745059341dc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a1a2ac924f0e4e4eb6e0c8d5f030bec3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 540.270165] env[61857]: INFO nova.compute.manager [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Rebuilding instance [ 540.413882] env[61857]: DEBUG nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 540.426151] env[61857]: DEBUG nova.compute.manager [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 540.430806] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c01d355-0eba-4258-8b96-b02657d64e7f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.591651] env[61857]: DEBUG nova.network.neutron [req-637b63a6-fc58-458f-9663-fd1f65913f31 req-269346f0-ed34-4b46-9205-4a7b3ede932e service nova] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.602233] env[61857]: DEBUG nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 540.627814] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c784e4-93f4-4b40-b41f-6cffb1d328b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.648041] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d4906c-9029-48ae-ab2b-57cb968a18e3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.695539] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Releasing lock "refresh_cache-90b0c4ad-902e-425d-b66b-51918e6476d9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.695736] env[61857]: DEBUG nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 540.696035] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 540.697568] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d8023ecc-acdc-4df8-ad46-9a48120dff8a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.702170] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550ac49f-38fe-41fa-9e16-5a90c4856575 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.715503] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76187965-133b-4a6b-a63d-46fe9bec2c7c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.740343] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc62f17-3789-4db5-8845-30ce80c1fb7a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.748985] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 90b0c4ad-902e-425d-b66b-51918e6476d9 could not be found. [ 540.749238] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 540.749390] env[61857]: INFO nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 540.749632] env[61857]: DEBUG oslo.service.loopingcall [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.750285] env[61857]: DEBUG nova.compute.manager [-] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 540.750345] env[61857]: DEBUG nova.network.neutron [-] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 540.759866] env[61857]: DEBUG nova.compute.provider_tree [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.839228] env[61857]: DEBUG nova.network.neutron [-] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.874026] env[61857]: ERROR nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04, please check neutron logs for more information. [ 540.874026] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.874026] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.874026] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.874026] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.874026] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.874026] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.874026] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.874026] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.874026] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 540.874026] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.874026] env[61857]: ERROR nova.compute.manager raise self.value [ 540.874026] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.874026] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.874026] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.874026] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.874516] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.874516] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.874516] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04, please check neutron logs for more information. [ 540.874516] env[61857]: ERROR nova.compute.manager [ 540.874516] env[61857]: Traceback (most recent call last): [ 540.874516] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.874516] env[61857]: listener.cb(fileno) [ 540.874516] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.874516] env[61857]: result = function(*args, **kwargs) [ 540.874516] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.874516] env[61857]: return func(*args, **kwargs) [ 540.874516] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.874516] env[61857]: raise e [ 540.874516] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.874516] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 540.874516] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.874516] env[61857]: created_port_ids = self._update_ports_for_instance( [ 540.874516] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.874516] env[61857]: with excutils.save_and_reraise_exception(): [ 540.874516] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.874516] env[61857]: self.force_reraise() [ 540.874516] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.874516] env[61857]: raise self.value [ 540.874516] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.874516] env[61857]: updated_port = self._update_port( [ 540.874516] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.874516] env[61857]: _ensure_no_port_binding_failure(port) [ 540.874516] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.874516] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.875273] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04, please check neutron logs for more information. [ 540.875273] env[61857]: Removing descriptor: 17 [ 540.875273] env[61857]: ERROR nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04, please check neutron logs for more information. [ 540.875273] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Traceback (most recent call last): [ 540.875273] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.875273] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] yield resources [ 540.875273] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.875273] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] self.driver.spawn(context, instance, image_meta, [ 540.875273] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 540.875273] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.875273] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.875273] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] vm_ref = self.build_virtual_machine(instance, [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] for vif in network_info: [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] return self._sync_wrapper(fn, *args, **kwargs) [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] self.wait() [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] self[:] = self._gt.wait() [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] return self._exit_event.wait() [ 540.877253] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] result = hub.switch() [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] return self.greenlet.switch() [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] result = function(*args, **kwargs) [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] return func(*args, **kwargs) [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] raise e [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] nwinfo = self.network_api.allocate_for_instance( [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.877599] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] created_port_ids = self._update_ports_for_instance( [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] with excutils.save_and_reraise_exception(): [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] self.force_reraise() [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] raise self.value [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] updated_port = self._update_port( [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] _ensure_no_port_binding_failure(port) [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.877923] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] raise exception.PortBindingFailed(port_id=port['id']) [ 540.878720] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] nova.exception.PortBindingFailed: Binding failed for port 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04, please check neutron logs for more information. [ 540.878720] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] [ 540.878720] env[61857]: INFO nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Terminating instance [ 540.879426] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Acquiring lock "refresh_cache-b34ca3cc-d26d-4068-a27d-a6d66404dfcb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.879426] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Acquired lock "refresh_cache-b34ca3cc-d26d-4068-a27d-a6d66404dfcb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.879569] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 540.900623] env[61857]: DEBUG nova.network.neutron [req-637b63a6-fc58-458f-9663-fd1f65913f31 req-269346f0-ed34-4b46-9205-4a7b3ede932e service nova] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.948520] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 540.952630] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b54ccf6a-ae19-497f-afcb-f30a79eaa0dd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.962472] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 540.962472] env[61857]: value = "task-950665" [ 540.962472] env[61857]: _type = "Task" [ 540.962472] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.978871] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.131632] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.265971] env[61857]: DEBUG nova.scheduler.client.report [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.349016] env[61857]: DEBUG nova.network.neutron [-] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.379807] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Successfully created port: 443c1a0b-6379-41eb-95ce-81de874f8587 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 541.406208] env[61857]: DEBUG oslo_concurrency.lockutils [req-637b63a6-fc58-458f-9663-fd1f65913f31 req-269346f0-ed34-4b46-9205-4a7b3ede932e service nova] Releasing lock "refresh_cache-61faaebe-374a-40ba-aefb-b199ea4dea93" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.413979] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquiring lock "2e7c154b-dab2-4384-bdb8-d098967654c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.413979] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lock "2e7c154b-dab2-4384-bdb8-d098967654c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.430694] env[61857]: DEBUG nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 541.443178] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.471738] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.472009] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.472202] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.472428] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.472633] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.472833] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.473158] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.473297] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.473501] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.473734] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.474458] env[61857]: DEBUG nova.virt.hardware [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.475621] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9146e8-9d25-42ae-9adc-dcf9bd58697e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.482247] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950665, 'name': PowerOffVM_Task, 'duration_secs': 0.123098} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.482914] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 541.486018] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 541.486018] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2243b054-91d0-4883-92c1-194f7f870cb2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.490740] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10be37ca-545a-458d-b976-9d5e4218bab3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.499890] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 541.500596] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-929a0f01-1b16-42e5-968c-b16faab1ea77 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.536944] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 541.536944] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 541.537150] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleting the datastore file [datastore2] bb418d62-d47c-42c5-908e-26f1eb034e46 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 541.537444] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-826fbba5-50ec-453c-b4c5-6c2b60d2feb7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.546979] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 541.546979] env[61857]: value = "task-950667" [ 541.546979] env[61857]: _type = "Task" [ 541.546979] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.558426] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.650638] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.771486] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.773069] env[61857]: DEBUG nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 541.774987] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.315s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.851934] env[61857]: INFO nova.compute.manager [-] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Took 1.10 seconds to deallocate network for instance. [ 541.852635] env[61857]: DEBUG nova.compute.claims [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 541.852823] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.915853] env[61857]: DEBUG nova.compute.manager [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.065149] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16983} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.065452] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 542.065638] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 542.065809] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 542.156549] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Releasing lock "refresh_cache-b34ca3cc-d26d-4068-a27d-a6d66404dfcb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.158168] env[61857]: DEBUG nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 542.158168] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 542.158168] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08b924e3-e34e-4388-9e4f-03bf509725a2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.169434] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df0bc923-e0f9-4e44-89df-a8529d5158a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.202977] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b34ca3cc-d26d-4068-a27d-a6d66404dfcb could not be found. [ 542.202977] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 542.202977] env[61857]: INFO nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 542.202977] env[61857]: DEBUG oslo.service.loopingcall [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.202977] env[61857]: DEBUG nova.compute.manager [-] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.202977] env[61857]: DEBUG nova.network.neutron [-] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 542.247617] env[61857]: DEBUG nova.network.neutron [-] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.280858] env[61857]: DEBUG nova.compute.utils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.287252] env[61857]: DEBUG nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 542.287606] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 542.358210] env[61857]: DEBUG nova.policy [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1496582f8474b14baf63b215a80e64e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3688a3ca85044527a84ad1a7b201fbdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 542.452993] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.487990] env[61857]: DEBUG nova.compute.manager [req-1552ab5e-335d-40f3-8c82-7a9bcdd41f10 req-d4d2c2b4-0c40-46e9-818d-9d37a76fea16 service nova] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Received event network-changed-808e5c82-8f76-4fda-b28c-9aaa6a1a5f04 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 542.488222] env[61857]: DEBUG nova.compute.manager [req-1552ab5e-335d-40f3-8c82-7a9bcdd41f10 req-d4d2c2b4-0c40-46e9-818d-9d37a76fea16 service nova] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Refreshing instance network info cache due to event network-changed-808e5c82-8f76-4fda-b28c-9aaa6a1a5f04. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 542.488445] env[61857]: DEBUG oslo_concurrency.lockutils [req-1552ab5e-335d-40f3-8c82-7a9bcdd41f10 req-d4d2c2b4-0c40-46e9-818d-9d37a76fea16 service nova] Acquiring lock "refresh_cache-b34ca3cc-d26d-4068-a27d-a6d66404dfcb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.488590] env[61857]: DEBUG oslo_concurrency.lockutils [req-1552ab5e-335d-40f3-8c82-7a9bcdd41f10 req-d4d2c2b4-0c40-46e9-818d-9d37a76fea16 service nova] Acquired lock "refresh_cache-b34ca3cc-d26d-4068-a27d-a6d66404dfcb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.488753] env[61857]: DEBUG nova.network.neutron [req-1552ab5e-335d-40f3-8c82-7a9bcdd41f10 req-d4d2c2b4-0c40-46e9-818d-9d37a76fea16 service nova] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Refreshing network info cache for port 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 542.585080] env[61857]: ERROR nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bba41198-8b1a-4895-ae13-fd458fd4eb5d, please check neutron logs for more information. [ 542.585080] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 542.585080] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.585080] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 542.585080] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.585080] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 542.585080] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.585080] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 542.585080] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.585080] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 542.585080] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.585080] env[61857]: ERROR nova.compute.manager raise self.value [ 542.585080] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.585080] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 542.585080] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.585080] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 542.585552] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.585552] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 542.585552] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bba41198-8b1a-4895-ae13-fd458fd4eb5d, please check neutron logs for more information. [ 542.585552] env[61857]: ERROR nova.compute.manager [ 542.585552] env[61857]: Traceback (most recent call last): [ 542.585552] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 542.585552] env[61857]: listener.cb(fileno) [ 542.585552] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.585552] env[61857]: result = function(*args, **kwargs) [ 542.585552] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 542.585552] env[61857]: return func(*args, **kwargs) [ 542.585552] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.585552] env[61857]: raise e [ 542.585552] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.585552] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 542.585552] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.585552] env[61857]: created_port_ids = self._update_ports_for_instance( [ 542.585552] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.585552] env[61857]: with excutils.save_and_reraise_exception(): [ 542.585552] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.585552] env[61857]: self.force_reraise() [ 542.585552] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.585552] env[61857]: raise self.value [ 542.585552] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.585552] env[61857]: updated_port = self._update_port( [ 542.585552] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.585552] env[61857]: _ensure_no_port_binding_failure(port) [ 542.585552] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.585552] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 542.589417] env[61857]: nova.exception.PortBindingFailed: Binding failed for port bba41198-8b1a-4895-ae13-fd458fd4eb5d, please check neutron logs for more information. [ 542.589417] env[61857]: Removing descriptor: 14 [ 542.589417] env[61857]: ERROR nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bba41198-8b1a-4895-ae13-fd458fd4eb5d, please check neutron logs for more information. [ 542.589417] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Traceback (most recent call last): [ 542.589417] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 542.589417] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] yield resources [ 542.589417] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 542.589417] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] self.driver.spawn(context, instance, image_meta, [ 542.589417] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 542.589417] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 542.589417] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 542.589417] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] vm_ref = self.build_virtual_machine(instance, [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] vif_infos = vmwarevif.get_vif_info(self._session, [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] for vif in network_info: [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] return self._sync_wrapper(fn, *args, **kwargs) [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] self.wait() [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] self[:] = self._gt.wait() [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] return self._exit_event.wait() [ 542.589735] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] result = hub.switch() [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] return self.greenlet.switch() [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] result = function(*args, **kwargs) [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] return func(*args, **kwargs) [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] raise e [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] nwinfo = self.network_api.allocate_for_instance( [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.590135] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] created_port_ids = self._update_ports_for_instance( [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] with excutils.save_and_reraise_exception(): [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] self.force_reraise() [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] raise self.value [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] updated_port = self._update_port( [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] _ensure_no_port_binding_failure(port) [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.590474] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] raise exception.PortBindingFailed(port_id=port['id']) [ 542.590769] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] nova.exception.PortBindingFailed: Binding failed for port bba41198-8b1a-4895-ae13-fd458fd4eb5d, please check neutron logs for more information. [ 542.590769] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] [ 542.590769] env[61857]: INFO nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Terminating instance [ 542.598617] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.598617] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.598617] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 542.659470] env[61857]: ERROR nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 27d6f88e-1abe-4552-b7a1-839802105483, please check neutron logs for more information. [ 542.659470] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 542.659470] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.659470] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 542.659470] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.659470] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 542.659470] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.659470] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 542.659470] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.659470] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 542.659470] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.659470] env[61857]: ERROR nova.compute.manager raise self.value [ 542.659470] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.659470] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 542.659470] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.659470] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 542.659943] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.659943] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 542.659943] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 27d6f88e-1abe-4552-b7a1-839802105483, please check neutron logs for more information. [ 542.659943] env[61857]: ERROR nova.compute.manager [ 542.659943] env[61857]: Traceback (most recent call last): [ 542.659943] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 542.659943] env[61857]: listener.cb(fileno) [ 542.659943] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.659943] env[61857]: result = function(*args, **kwargs) [ 542.659943] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 542.659943] env[61857]: return func(*args, **kwargs) [ 542.659943] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.659943] env[61857]: raise e [ 542.659943] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.659943] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 542.659943] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.659943] env[61857]: created_port_ids = self._update_ports_for_instance( [ 542.659943] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.659943] env[61857]: with excutils.save_and_reraise_exception(): [ 542.659943] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.659943] env[61857]: self.force_reraise() [ 542.659943] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.659943] env[61857]: raise self.value [ 542.659943] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.659943] env[61857]: updated_port = self._update_port( [ 542.659943] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.659943] env[61857]: _ensure_no_port_binding_failure(port) [ 542.659943] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.659943] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 542.660838] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 27d6f88e-1abe-4552-b7a1-839802105483, please check neutron logs for more information. [ 542.660838] env[61857]: Removing descriptor: 18 [ 542.660838] env[61857]: ERROR nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 27d6f88e-1abe-4552-b7a1-839802105483, please check neutron logs for more information. [ 542.660838] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Traceback (most recent call last): [ 542.660838] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 542.660838] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] yield resources [ 542.660838] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 542.660838] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] self.driver.spawn(context, instance, image_meta, [ 542.660838] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 542.660838] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 542.660838] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 542.660838] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] vm_ref = self.build_virtual_machine(instance, [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] vif_infos = vmwarevif.get_vif_info(self._session, [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] for vif in network_info: [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] return self._sync_wrapper(fn, *args, **kwargs) [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] self.wait() [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] self[:] = self._gt.wait() [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] return self._exit_event.wait() [ 542.661166] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] result = hub.switch() [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] return self.greenlet.switch() [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] result = function(*args, **kwargs) [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] return func(*args, **kwargs) [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] raise e [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] nwinfo = self.network_api.allocate_for_instance( [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.661516] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] created_port_ids = self._update_ports_for_instance( [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] with excutils.save_and_reraise_exception(): [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] self.force_reraise() [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] raise self.value [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] updated_port = self._update_port( [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] _ensure_no_port_binding_failure(port) [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.661869] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] raise exception.PortBindingFailed(port_id=port['id']) [ 542.662207] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] nova.exception.PortBindingFailed: Binding failed for port 27d6f88e-1abe-4552-b7a1-839802105483, please check neutron logs for more information. [ 542.662207] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] [ 542.662207] env[61857]: INFO nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Terminating instance [ 542.663875] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Acquiring lock "refresh_cache-6acac8fe-a2a7-426e-912a-0292d351887e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.664035] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Acquired lock "refresh_cache-6acac8fe-a2a7-426e-912a-0292d351887e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.664201] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 542.754643] env[61857]: DEBUG nova.network.neutron [-] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.793175] env[61857]: DEBUG nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 542.827476] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance bb418d62-d47c-42c5-908e-26f1eb034e46 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.827621] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 61faaebe-374a-40ba-aefb-b199ea4dea93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.827753] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 90b0c4ad-902e-425d-b66b-51918e6476d9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.827870] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance b34ca3cc-d26d-4068-a27d-a6d66404dfcb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.827984] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 6acac8fe-a2a7-426e-912a-0292d351887e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.828110] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance b0fd68df-fcba-494d-bf79-a0cf2ea82ed5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.828222] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 4c7dab9e-4527-46b6-a1ec-6e6935768b75 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.828684] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 67ff728e-37b1-4633-bfd2-75fe4b1c7d13 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.969969] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Successfully created port: d9602eb9-b5bf-4cf1-a971-2ca9191b94d9 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 543.042147] env[61857]: DEBUG nova.network.neutron [req-1552ab5e-335d-40f3-8c82-7a9bcdd41f10 req-d4d2c2b4-0c40-46e9-818d-9d37a76fea16 service nova] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.115330] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.115570] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.115725] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.115904] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.116095] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.116210] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.116418] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.116661] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.116725] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.116886] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.117230] env[61857]: DEBUG nova.virt.hardware [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.119775] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a731a363-5c33-498d-ac35-206951d91762 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.127441] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b30e9c1-b923-4d63-957b-19a2ccb2c579 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.132395] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.145646] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 543.151629] env[61857]: DEBUG oslo.service.loopingcall [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 543.151869] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 543.152127] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-189ac6a6-8b07-4ecd-879e-82dd794d7fe5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.170633] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 543.170633] env[61857]: value = "task-950668" [ 543.170633] env[61857]: _type = "Task" [ 543.170633] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.180618] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950668, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.211011] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.259897] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.261280] env[61857]: DEBUG nova.network.neutron [req-1552ab5e-335d-40f3-8c82-7a9bcdd41f10 req-d4d2c2b4-0c40-46e9-818d-9d37a76fea16 service nova] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.264886] env[61857]: INFO nova.compute.manager [-] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Took 1.06 seconds to deallocate network for instance. [ 543.269839] env[61857]: DEBUG nova.compute.claims [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 543.270330] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.331957] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ad56d32b-c78d-45db-872a-1c4ea91ab909 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 543.411306] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.687901] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950668, 'name': CreateVM_Task, 'duration_secs': 0.290275} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.689905] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 543.691031] env[61857]: DEBUG oslo_vmware.service [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd497793-f8ef-4428-9c01-49331a26cce6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.703866] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.704121] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.704918] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 543.705244] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5d3dcb5-595e-44b9-aada-6497b2635861 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.713898] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 543.713898] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52801ace-8201-a16f-6e3c-dec81831f52d" [ 543.713898] env[61857]: _type = "Task" [ 543.713898] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.723370] env[61857]: DEBUG nova.compute.manager [req-6151b267-1301-44f6-9e48-b46fb7a7076a req-95ab8a19-6b64-4d72-a60e-73cbc185d02e service nova] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Received event network-vif-deleted-ea2b2110-9959-44fa-a150-e11ae8904d4c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 543.723370] env[61857]: DEBUG nova.compute.manager [req-6151b267-1301-44f6-9e48-b46fb7a7076a req-95ab8a19-6b64-4d72-a60e-73cbc185d02e service nova] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Received event network-changed-4b0da35d-33e2-494e-a2da-473d4e03a462 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 543.723370] env[61857]: DEBUG nova.compute.manager [req-6151b267-1301-44f6-9e48-b46fb7a7076a req-95ab8a19-6b64-4d72-a60e-73cbc185d02e service nova] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Refreshing instance network info cache due to event network-changed-4b0da35d-33e2-494e-a2da-473d4e03a462. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 543.723370] env[61857]: DEBUG oslo_concurrency.lockutils [req-6151b267-1301-44f6-9e48-b46fb7a7076a req-95ab8a19-6b64-4d72-a60e-73cbc185d02e service nova] Acquiring lock "refresh_cache-90b0c4ad-902e-425d-b66b-51918e6476d9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.723370] env[61857]: DEBUG oslo_concurrency.lockutils [req-6151b267-1301-44f6-9e48-b46fb7a7076a req-95ab8a19-6b64-4d72-a60e-73cbc185d02e service nova] Acquired lock "refresh_cache-90b0c4ad-902e-425d-b66b-51918e6476d9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.723530] env[61857]: DEBUG nova.network.neutron [req-6151b267-1301-44f6-9e48-b46fb7a7076a req-95ab8a19-6b64-4d72-a60e-73cbc185d02e service nova] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Refreshing network info cache for port 4b0da35d-33e2-494e-a2da-473d4e03a462 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 543.733138] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.736259] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 543.736259] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.736259] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.736259] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 543.736259] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b2669da-f23f-4235-a5ab-062a8c8ad1cb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.747305] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 543.747305] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 543.748217] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6501ed4b-8a53-493c-9288-47c299bfd283 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.757482] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-094cf830-b340-4ff8-ac43-9036e1dfe272 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.764617] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 543.764617] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524e77c9-2b1f-614b-a2cf-f60f1f4b92d2" [ 543.764617] env[61857]: _type = "Task" [ 543.764617] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.768165] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.768637] env[61857]: DEBUG nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 543.768858] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 543.769414] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f82af107-17c5-48b8-98cf-13b9a6994e65 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.773610] env[61857]: DEBUG oslo_concurrency.lockutils [req-1552ab5e-335d-40f3-8c82-7a9bcdd41f10 req-d4d2c2b4-0c40-46e9-818d-9d37a76fea16 service nova] Releasing lock "refresh_cache-b34ca3cc-d26d-4068-a27d-a6d66404dfcb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.779254] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Preparing fetch location {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 543.779254] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Creating directory with path [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 543.779254] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a6075e0-9fad-4db7-b549-e7aa9ad09c8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.786856] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9abcbd-4319-488b-8c90-c633abfbe003 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.807751] env[61857]: DEBUG nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 543.825230] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b0fd68df-fcba-494d-bf79-a0cf2ea82ed5 could not be found. [ 543.825532] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 543.826347] env[61857]: INFO nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Took 0.06 seconds to destroy the instance on the hypervisor. [ 543.826347] env[61857]: DEBUG oslo.service.loopingcall [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 543.827634] env[61857]: DEBUG nova.compute.manager [-] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 543.827634] env[61857]: DEBUG nova.network.neutron [-] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 543.829771] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Created directory with path [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 543.829771] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Fetch image to [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 543.829847] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Downloading image file data 2e0c8bb4-0827-486c-98ce-7994555ffc23 to [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk on the data store datastore1 {{(pid=61857) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 543.830782] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9f9d23-549d-48e0-9862-0e3cdf08d15d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.837968] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 2e7c154b-dab2-4384-bdb8-d098967654c5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 543.838215] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 543.838360] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 543.841765] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16ef192-9512-4644-a153-1e7ee4d04612 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.859722] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91cedc66-eb44-4934-9e36-287cb795533e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.872578] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.872818] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.872969] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.873192] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.873314] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.873463] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.874056] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.874056] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.874056] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.875649] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.875649] env[61857]: DEBUG nova.virt.hardware [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.876339] env[61857]: DEBUG nova.network.neutron [-] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.882026] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dea8fa5-7899-4163-ba8c-30a602565cb5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.922180] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff695af-4232-458c-b53b-0f9325f473a2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.927486] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Releasing lock "refresh_cache-6acac8fe-a2a7-426e-912a-0292d351887e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.927617] env[61857]: DEBUG nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 543.927956] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 543.931098] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-018f3282-ee17-4e51-890e-49b13e5534ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.933437] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0e5424-626e-415b-8da2-482de323ecdd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.952676] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709ad332-b6f2-4c50-918f-78e7d3274f96 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.968457] env[61857]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-abb3af35-6cf9-4928-899e-8b03bb56521a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.978562] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6acac8fe-a2a7-426e-912a-0292d351887e could not be found. [ 543.978782] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 543.978886] env[61857]: INFO nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 543.981026] env[61857]: DEBUG oslo.service.loopingcall [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 543.981026] env[61857]: DEBUG nova.compute.manager [-] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 543.981026] env[61857]: DEBUG nova.network.neutron [-] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 543.992068] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Downloading image file data 2e0c8bb4-0827-486c-98ce-7994555ffc23 to the data store datastore1 {{(pid=61857) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 543.998585] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquiring lock "e8c59da6-c226-4c83-869b-6a200cf1c0b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.999377] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "e8c59da6-c226-4c83-869b-6a200cf1c0b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.097376] env[61857]: DEBUG oslo_vmware.rw_handles [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61857) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 544.173892] env[61857]: DEBUG nova.network.neutron [-] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.201813] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29bfdea-e958-481b-8267-b83e5fe0d4ee {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.214316] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b195290-43bc-4e4e-b028-bc34abe7993f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.259465] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e281c6-6f9a-48ff-99f3-f3a5cb7eb563 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.269754] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27b8324-8757-4105-8f95-076b5b83a994 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.288953] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.307169] env[61857]: DEBUG nova.network.neutron [req-6151b267-1301-44f6-9e48-b46fb7a7076a req-95ab8a19-6b64-4d72-a60e-73cbc185d02e service nova] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.349606] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Acquiring lock "f38e8137-da78-40f7-88a1-0f870a83ffd9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.349872] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Lock "f38e8137-da78-40f7-88a1-0f870a83ffd9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.389181] env[61857]: DEBUG nova.network.neutron [-] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.460504] env[61857]: DEBUG nova.network.neutron [req-6151b267-1301-44f6-9e48-b46fb7a7076a req-95ab8a19-6b64-4d72-a60e-73cbc185d02e service nova] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.504364] env[61857]: DEBUG nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.678074] env[61857]: DEBUG nova.network.neutron [-] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.727069] env[61857]: ERROR nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d9602eb9-b5bf-4cf1-a971-2ca9191b94d9, please check neutron logs for more information. [ 544.727069] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 544.727069] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.727069] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 544.727069] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.727069] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 544.727069] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.727069] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 544.727069] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.727069] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 544.727069] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.727069] env[61857]: ERROR nova.compute.manager raise self.value [ 544.727069] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.727069] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 544.727069] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.727069] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 544.727598] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.727598] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 544.727598] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d9602eb9-b5bf-4cf1-a971-2ca9191b94d9, please check neutron logs for more information. [ 544.727598] env[61857]: ERROR nova.compute.manager [ 544.727598] env[61857]: Traceback (most recent call last): [ 544.727598] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 544.727598] env[61857]: listener.cb(fileno) [ 544.727598] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.727598] env[61857]: result = function(*args, **kwargs) [ 544.727598] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.727598] env[61857]: return func(*args, **kwargs) [ 544.727598] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.727598] env[61857]: raise e [ 544.727598] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.727598] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 544.727598] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.727598] env[61857]: created_port_ids = self._update_ports_for_instance( [ 544.727598] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.727598] env[61857]: with excutils.save_and_reraise_exception(): [ 544.727598] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.727598] env[61857]: self.force_reraise() [ 544.727598] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.727598] env[61857]: raise self.value [ 544.727598] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.727598] env[61857]: updated_port = self._update_port( [ 544.727598] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.727598] env[61857]: _ensure_no_port_binding_failure(port) [ 544.727598] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.727598] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 544.728303] env[61857]: nova.exception.PortBindingFailed: Binding failed for port d9602eb9-b5bf-4cf1-a971-2ca9191b94d9, please check neutron logs for more information. [ 544.728303] env[61857]: Removing descriptor: 16 [ 544.728303] env[61857]: ERROR nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d9602eb9-b5bf-4cf1-a971-2ca9191b94d9, please check neutron logs for more information. [ 544.728303] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Traceback (most recent call last): [ 544.728303] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 544.728303] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] yield resources [ 544.728303] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.728303] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] self.driver.spawn(context, instance, image_meta, [ 544.728303] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 544.728303] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.728303] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.728303] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] vm_ref = self.build_virtual_machine(instance, [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] for vif in network_info: [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] return self._sync_wrapper(fn, *args, **kwargs) [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] self.wait() [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] self[:] = self._gt.wait() [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] return self._exit_event.wait() [ 544.728685] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] result = hub.switch() [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] return self.greenlet.switch() [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] result = function(*args, **kwargs) [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] return func(*args, **kwargs) [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] raise e [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] nwinfo = self.network_api.allocate_for_instance( [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.729029] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] created_port_ids = self._update_ports_for_instance( [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] with excutils.save_and_reraise_exception(): [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] self.force_reraise() [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] raise self.value [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] updated_port = self._update_port( [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] _ensure_no_port_binding_failure(port) [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.729524] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] raise exception.PortBindingFailed(port_id=port['id']) [ 544.729833] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] nova.exception.PortBindingFailed: Binding failed for port d9602eb9-b5bf-4cf1-a971-2ca9191b94d9, please check neutron logs for more information. [ 544.729833] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] [ 544.729833] env[61857]: INFO nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Terminating instance [ 544.731361] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "refresh_cache-67ff728e-37b1-4633-bfd2-75fe4b1c7d13" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.731560] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "refresh_cache-67ff728e-37b1-4633-bfd2-75fe4b1c7d13" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.731761] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.794111] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 544.839998] env[61857]: DEBUG oslo_vmware.rw_handles [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Completed reading data from the image iterator. {{(pid=61857) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 544.840353] env[61857]: DEBUG oslo_vmware.rw_handles [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 544.891606] env[61857]: INFO nova.compute.manager [-] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Took 1.06 seconds to deallocate network for instance. [ 544.894219] env[61857]: DEBUG nova.compute.claims [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 544.894344] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.898554] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Downloaded image file data 2e0c8bb4-0827-486c-98ce-7994555ffc23 to vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk on the data store datastore1 {{(pid=61857) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 544.900280] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Caching image {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 544.900514] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Copying Virtual Disk [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk to [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 544.901241] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d37e79d-accf-4333-9d31-b6e8d4098a57 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.909279] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 544.909279] env[61857]: value = "task-950669" [ 544.909279] env[61857]: _type = "Task" [ 544.909279] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.913333] env[61857]: ERROR nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 443c1a0b-6379-41eb-95ce-81de874f8587, please check neutron logs for more information. [ 544.913333] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 544.913333] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.913333] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 544.913333] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.913333] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 544.913333] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.913333] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 544.913333] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.913333] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 544.913333] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.913333] env[61857]: ERROR nova.compute.manager raise self.value [ 544.913333] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.913333] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 544.913333] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.913333] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 544.913903] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.913903] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 544.913903] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 443c1a0b-6379-41eb-95ce-81de874f8587, please check neutron logs for more information. [ 544.913903] env[61857]: ERROR nova.compute.manager [ 544.913903] env[61857]: Traceback (most recent call last): [ 544.913903] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 544.913903] env[61857]: listener.cb(fileno) [ 544.913903] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.913903] env[61857]: result = function(*args, **kwargs) [ 544.913903] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.913903] env[61857]: return func(*args, **kwargs) [ 544.913903] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.913903] env[61857]: raise e [ 544.913903] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.913903] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 544.913903] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.913903] env[61857]: created_port_ids = self._update_ports_for_instance( [ 544.913903] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.913903] env[61857]: with excutils.save_and_reraise_exception(): [ 544.913903] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.913903] env[61857]: self.force_reraise() [ 544.913903] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.913903] env[61857]: raise self.value [ 544.913903] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.913903] env[61857]: updated_port = self._update_port( [ 544.913903] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.913903] env[61857]: _ensure_no_port_binding_failure(port) [ 544.913903] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.913903] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 544.914683] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 443c1a0b-6379-41eb-95ce-81de874f8587, please check neutron logs for more information. [ 544.914683] env[61857]: Removing descriptor: 20 [ 544.914683] env[61857]: ERROR nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 443c1a0b-6379-41eb-95ce-81de874f8587, please check neutron logs for more information. [ 544.914683] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Traceback (most recent call last): [ 544.914683] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 544.914683] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] yield resources [ 544.914683] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.914683] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] self.driver.spawn(context, instance, image_meta, [ 544.914683] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 544.914683] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.914683] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.914683] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] vm_ref = self.build_virtual_machine(instance, [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] for vif in network_info: [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] return self._sync_wrapper(fn, *args, **kwargs) [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] self.wait() [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] self[:] = self._gt.wait() [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] return self._exit_event.wait() [ 544.915068] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] result = hub.switch() [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] return self.greenlet.switch() [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] result = function(*args, **kwargs) [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] return func(*args, **kwargs) [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] raise e [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] nwinfo = self.network_api.allocate_for_instance( [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.915500] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] created_port_ids = self._update_ports_for_instance( [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] with excutils.save_and_reraise_exception(): [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] self.force_reraise() [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] raise self.value [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] updated_port = self._update_port( [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] _ensure_no_port_binding_failure(port) [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.915894] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] raise exception.PortBindingFailed(port_id=port['id']) [ 544.916272] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] nova.exception.PortBindingFailed: Binding failed for port 443c1a0b-6379-41eb-95ce-81de874f8587, please check neutron logs for more information. [ 544.916272] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] [ 544.916272] env[61857]: INFO nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Terminating instance [ 544.921829] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950669, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.922090] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Acquiring lock "refresh_cache-4c7dab9e-4527-46b6-a1ec-6e6935768b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.922090] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Acquired lock "refresh_cache-4c7dab9e-4527-46b6-a1ec-6e6935768b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.922256] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.964342] env[61857]: DEBUG oslo_concurrency.lockutils [req-6151b267-1301-44f6-9e48-b46fb7a7076a req-95ab8a19-6b64-4d72-a60e-73cbc185d02e service nova] Releasing lock "refresh_cache-90b0c4ad-902e-425d-b66b-51918e6476d9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.964342] env[61857]: DEBUG nova.compute.manager [req-6151b267-1301-44f6-9e48-b46fb7a7076a req-95ab8a19-6b64-4d72-a60e-73cbc185d02e service nova] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Received event network-vif-deleted-4b0da35d-33e2-494e-a2da-473d4e03a462 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 545.029052] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.180542] env[61857]: INFO nova.compute.manager [-] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Took 1.20 seconds to deallocate network for instance. [ 545.185376] env[61857]: DEBUG nova.compute.claims [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 545.185376] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.284039] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.301113] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 545.301113] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.524s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.301113] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.569s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.420052] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950669, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.449997] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.581360] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.615856] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.921815] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950669, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651521} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.923822] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Copied Virtual Disk [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk to [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 545.924554] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleting the datastore file [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23/tmp-sparse.vmdk {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 545.924951] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-747d0919-b410-4c66-a047-60075c55731a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.934285] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 545.934285] env[61857]: value = "task-950670" [ 545.934285] env[61857]: _type = "Task" [ 545.934285] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.954615] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.087036] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Releasing lock "refresh_cache-4c7dab9e-4527-46b6-a1ec-6e6935768b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.087036] env[61857]: DEBUG nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 546.087036] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 546.087036] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41a4121a-2f03-40d2-a9e7-9df771b6f272 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.099464] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e604d65-6268-4f04-80f1-4fd9a35ac438 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.126418] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "refresh_cache-67ff728e-37b1-4633-bfd2-75fe4b1c7d13" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.127451] env[61857]: DEBUG nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 546.127451] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 546.128797] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c7dab9e-4527-46b6-a1ec-6e6935768b75 could not be found. [ 546.129027] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 546.129306] env[61857]: INFO nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Took 0.04 seconds to destroy the instance on the hypervisor. [ 546.130273] env[61857]: DEBUG oslo.service.loopingcall [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 546.130273] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f8494f8a-23db-4cc9-bbac-6cdc329f861a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.132864] env[61857]: DEBUG nova.compute.manager [-] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.132946] env[61857]: DEBUG nova.network.neutron [-] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 546.138397] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8ef77c-9f0c-4611-adfd-9dd4a6705acc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.145253] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88716a34-7ed2-4ee9-b818-d1eb8048d938 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.162139] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad75b9f7-52d0-4f51-b0b7-3542de4c695d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.167921] env[61857]: DEBUG nova.compute.manager [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Received event network-vif-deleted-808e5c82-8f76-4fda-b28c-9aaa6a1a5f04 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 546.168697] env[61857]: DEBUG nova.compute.manager [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Received event network-changed-27d6f88e-1abe-4552-b7a1-839802105483 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 546.169247] env[61857]: DEBUG nova.compute.manager [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Refreshing instance network info cache due to event network-changed-27d6f88e-1abe-4552-b7a1-839802105483. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 546.169877] env[61857]: DEBUG oslo_concurrency.lockutils [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] Acquiring lock "refresh_cache-6acac8fe-a2a7-426e-912a-0292d351887e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.170203] env[61857]: DEBUG oslo_concurrency.lockutils [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] Acquired lock "refresh_cache-6acac8fe-a2a7-426e-912a-0292d351887e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.170395] env[61857]: DEBUG nova.network.neutron [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Refreshing network info cache for port 27d6f88e-1abe-4552-b7a1-839802105483 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 546.177624] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 67ff728e-37b1-4633-bfd2-75fe4b1c7d13 could not be found. [ 546.177844] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 546.178092] env[61857]: INFO nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Took 0.05 seconds to destroy the instance on the hypervisor. [ 546.178382] env[61857]: DEBUG oslo.service.loopingcall [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 546.179193] env[61857]: DEBUG nova.compute.manager [-] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.179284] env[61857]: DEBUG nova.network.neutron [-] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 546.211383] env[61857]: DEBUG nova.network.neutron [-] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.215688] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db895b8-73a3-4ece-ba4c-446b84a6be4d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.219365] env[61857]: DEBUG nova.network.neutron [-] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.221111] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquiring lock "0b6029b8-3267-4662-a1af-8e0b6af3288f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.221445] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "0b6029b8-3267-4662-a1af-8e0b6af3288f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.229188] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6fb972-a65b-44ed-95fb-2ac04012c4c3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.245308] env[61857]: DEBUG nova.compute.provider_tree [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.445861] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02248} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.448016] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 546.448016] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Moving file from [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21/2e0c8bb4-0827-486c-98ce-7994555ffc23 to [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23. {{(pid=61857) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 546.448016] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-dfd0db3e-2f70-417a-98bb-49b0d609e31f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.454048] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 546.454048] env[61857]: value = "task-950671" [ 546.454048] env[61857]: _type = "Task" [ 546.454048] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.466427] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950671, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.711942] env[61857]: DEBUG nova.network.neutron [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.723015] env[61857]: DEBUG nova.network.neutron [-] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.724489] env[61857]: DEBUG nova.network.neutron [-] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.748927] env[61857]: DEBUG nova.scheduler.client.report [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.911419] env[61857]: DEBUG nova.network.neutron [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.968015] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950671, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.027672} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.968015] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] File moved {{(pid=61857) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 546.969699] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Cleaning up location [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 546.969699] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleting the datastore file [datastore1] vmware_temp/240bc4b9-5e63-4ef4-84cf-ae78a4e93e21 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 546.969699] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc455fe3-79df-4ecc-ab9a-9684529d040b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.978459] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 546.978459] env[61857]: value = "task-950672" [ 546.978459] env[61857]: _type = "Task" [ 546.978459] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.993332] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950672, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.047027] env[61857]: DEBUG nova.compute.manager [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Received event network-changed-bba41198-8b1a-4895-ae13-fd458fd4eb5d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 547.047372] env[61857]: DEBUG nova.compute.manager [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Refreshing instance network info cache due to event network-changed-bba41198-8b1a-4895-ae13-fd458fd4eb5d. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 547.048359] env[61857]: DEBUG oslo_concurrency.lockutils [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] Acquiring lock "refresh_cache-b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.048359] env[61857]: DEBUG oslo_concurrency.lockutils [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] Acquired lock "refresh_cache-b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.048359] env[61857]: DEBUG nova.network.neutron [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Refreshing network info cache for port bba41198-8b1a-4895-ae13-fd458fd4eb5d {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 547.225367] env[61857]: INFO nova.compute.manager [-] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Took 1.09 seconds to deallocate network for instance. [ 547.228379] env[61857]: INFO nova.compute.manager [-] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Took 1.05 seconds to deallocate network for instance. [ 547.228905] env[61857]: DEBUG nova.compute.claims [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 547.229547] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.232462] env[61857]: DEBUG nova.compute.claims [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 547.232635] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.262566] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.263592] env[61857]: ERROR nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ea2b2110-9959-44fa-a150-e11ae8904d4c, please check neutron logs for more information. [ 547.263592] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Traceback (most recent call last): [ 547.263592] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.263592] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] self.driver.spawn(context, instance, image_meta, [ 547.263592] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 547.263592] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.263592] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.263592] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] vm_ref = self.build_virtual_machine(instance, [ 547.263592] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.263592] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.263592] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] for vif in network_info: [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] return self._sync_wrapper(fn, *args, **kwargs) [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] self.wait() [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] self[:] = self._gt.wait() [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] return self._exit_event.wait() [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] result = hub.switch() [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.263992] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] return self.greenlet.switch() [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] result = function(*args, **kwargs) [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] return func(*args, **kwargs) [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] raise e [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] nwinfo = self.network_api.allocate_for_instance( [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] created_port_ids = self._update_ports_for_instance( [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] with excutils.save_and_reraise_exception(): [ 547.264320] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] self.force_reraise() [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] raise self.value [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] updated_port = self._update_port( [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] _ensure_no_port_binding_failure(port) [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] raise exception.PortBindingFailed(port_id=port['id']) [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] nova.exception.PortBindingFailed: Binding failed for port ea2b2110-9959-44fa-a150-e11ae8904d4c, please check neutron logs for more information. [ 547.264734] env[61857]: ERROR nova.compute.manager [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] [ 547.265030] env[61857]: DEBUG nova.compute.utils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Binding failed for port ea2b2110-9959-44fa-a150-e11ae8904d4c, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.266132] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.135s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.267644] env[61857]: INFO nova.compute.claims [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.272249] env[61857]: DEBUG nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Build of instance 61faaebe-374a-40ba-aefb-b199ea4dea93 was re-scheduled: Binding failed for port ea2b2110-9959-44fa-a150-e11ae8904d4c, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 547.272707] env[61857]: DEBUG nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 547.272944] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquiring lock "refresh_cache-61faaebe-374a-40ba-aefb-b199ea4dea93" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.273099] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquired lock "refresh_cache-61faaebe-374a-40ba-aefb-b199ea4dea93" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.273258] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 547.414347] env[61857]: DEBUG oslo_concurrency.lockutils [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] Releasing lock "refresh_cache-6acac8fe-a2a7-426e-912a-0292d351887e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.414736] env[61857]: DEBUG nova.compute.manager [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Received event network-vif-deleted-27d6f88e-1abe-4552-b7a1-839802105483 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 547.414821] env[61857]: DEBUG nova.compute.manager [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Received event network-changed-443c1a0b-6379-41eb-95ce-81de874f8587 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 547.414940] env[61857]: DEBUG nova.compute.manager [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Refreshing instance network info cache due to event network-changed-443c1a0b-6379-41eb-95ce-81de874f8587. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 547.415164] env[61857]: DEBUG oslo_concurrency.lockutils [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] Acquiring lock "refresh_cache-4c7dab9e-4527-46b6-a1ec-6e6935768b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.415719] env[61857]: DEBUG oslo_concurrency.lockutils [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] Acquired lock "refresh_cache-4c7dab9e-4527-46b6-a1ec-6e6935768b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.415719] env[61857]: DEBUG nova.network.neutron [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Refreshing network info cache for port 443c1a0b-6379-41eb-95ce-81de874f8587 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 547.493514] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950672, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02548} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.493514] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 547.493514] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-559d714c-14b4-4384-8b89-fe060d84d19e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.501346] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 547.501346] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525e299d-bfbd-c4e6-db6e-7ecfa2d43e18" [ 547.501346] env[61857]: _type = "Task" [ 547.501346] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.510522] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525e299d-bfbd-c4e6-db6e-7ecfa2d43e18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.591414] env[61857]: DEBUG nova.network.neutron [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.822366] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.925623] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.951373] env[61857]: DEBUG nova.network.neutron [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.983875] env[61857]: DEBUG nova.network.neutron [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.015581] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525e299d-bfbd-c4e6-db6e-7ecfa2d43e18, 'name': SearchDatastore_Task, 'duration_secs': 0.008866} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.015837] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.016101] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 548.017634] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43378dc8-02ce-4afc-aec4-1c3b85a4cb32 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.027133] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 548.027133] env[61857]: value = "task-950673" [ 548.027133] env[61857]: _type = "Task" [ 548.027133] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.037321] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950673, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.045231] env[61857]: DEBUG nova.network.neutron [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.428499] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Releasing lock "refresh_cache-61faaebe-374a-40ba-aefb-b199ea4dea93" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.428752] env[61857]: DEBUG nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 548.428963] env[61857]: DEBUG nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.429653] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 548.485032] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.489475] env[61857]: DEBUG oslo_concurrency.lockutils [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] Releasing lock "refresh_cache-b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.489736] env[61857]: DEBUG nova.compute.manager [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Received event network-vif-deleted-bba41198-8b1a-4895-ae13-fd458fd4eb5d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 548.489913] env[61857]: DEBUG nova.compute.manager [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Received event network-changed-d9602eb9-b5bf-4cf1-a971-2ca9191b94d9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 548.491371] env[61857]: DEBUG nova.compute.manager [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Refreshing instance network info cache due to event network-changed-d9602eb9-b5bf-4cf1-a971-2ca9191b94d9. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 548.491371] env[61857]: DEBUG oslo_concurrency.lockutils [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] Acquiring lock "refresh_cache-67ff728e-37b1-4633-bfd2-75fe4b1c7d13" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.491371] env[61857]: DEBUG oslo_concurrency.lockutils [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] Acquired lock "refresh_cache-67ff728e-37b1-4633-bfd2-75fe4b1c7d13" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.491371] env[61857]: DEBUG nova.network.neutron [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Refreshing network info cache for port d9602eb9-b5bf-4cf1-a971-2ca9191b94d9 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 548.538871] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950673, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447902} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.539425] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 548.539425] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 548.539622] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-89ea29f7-9187-462e-8a1e-969efab66bcc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.547588] env[61857]: DEBUG oslo_concurrency.lockutils [req-cede02d6-2ed8-4365-9755-9b4076685b57 req-8d82a2b4-76d1-4879-81dc-e15964a0b5a6 service nova] Releasing lock "refresh_cache-4c7dab9e-4527-46b6-a1ec-6e6935768b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.548517] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153a4f24-f365-484e-85c7-ca49b9cbe04b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.553037] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 548.553037] env[61857]: value = "task-950674" [ 548.553037] env[61857]: _type = "Task" [ 548.553037] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.559816] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642807aa-5b51-4f3b-a071-28ad232966fb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.566281] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950674, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.599965] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2cab0d-b7cf-49ac-92ab-2ec4883e80c9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.608747] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c77bf0-f2d2-451e-a73f-4533ac059698 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.627708] env[61857]: DEBUG nova.compute.provider_tree [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.988644] env[61857]: DEBUG nova.network.neutron [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.047671] env[61857]: DEBUG nova.network.neutron [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.069738] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950674, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068842} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.070160] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 549.072914] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b451d8f-ba89-4aba-b459-4dc362214a90 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.100957] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 549.101361] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c81bce3-975b-4a08-a474-2beb021f5538 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.122978] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 549.122978] env[61857]: value = "task-950675" [ 549.122978] env[61857]: _type = "Task" [ 549.122978] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.133515] env[61857]: DEBUG nova.scheduler.client.report [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.138643] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950675, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.265386] env[61857]: DEBUG nova.network.neutron [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.455794] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Acquiring lock "e7e82254-4088-4232-b861-9aee864c3737" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.456065] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Lock "e7e82254-4088-4232-b861-9aee864c3737" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.496019] env[61857]: INFO nova.compute.manager [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: 61faaebe-374a-40ba-aefb-b199ea4dea93] Took 1.06 seconds to deallocate network for instance. [ 549.636649] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950675, 'name': ReconfigVM_Task, 'duration_secs': 0.327067} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.637790] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Reconfigured VM instance instance-00000001 to attach disk [datastore1] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 549.638829] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a855e63-0433-4762-ac14-bf27121ef996 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.643804] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.643804] env[61857]: DEBUG nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 549.647214] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.794s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.655727] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 549.655727] env[61857]: value = "task-950676" [ 549.655727] env[61857]: _type = "Task" [ 549.655727] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.668326] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950676, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.768523] env[61857]: DEBUG oslo_concurrency.lockutils [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] Releasing lock "refresh_cache-67ff728e-37b1-4633-bfd2-75fe4b1c7d13" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.768815] env[61857]: DEBUG nova.compute.manager [req-05eb79d0-b460-4684-bd0a-0fea9f8b058c req-1af5c7af-6bcb-4116-be6a-45cb28079af9 service nova] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Received event network-vif-deleted-d9602eb9-b5bf-4cf1-a971-2ca9191b94d9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 549.810274] env[61857]: DEBUG nova.compute.manager [req-34aeaee7-5be5-4201-a82f-76d76b15407b req-88433afe-2ea7-4889-b4fb-fb2ae2cb6e2c service nova] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Received event network-vif-deleted-443c1a0b-6379-41eb-95ce-81de874f8587 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 550.161647] env[61857]: DEBUG nova.compute.utils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.164333] env[61857]: DEBUG nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 550.164333] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 550.179093] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950676, 'name': Rename_Task, 'duration_secs': 0.159892} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.179396] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 550.179642] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c49e99b4-699e-44ef-ae61-71fad2662e8c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.189350] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 550.189350] env[61857]: value = "task-950677" [ 550.189350] env[61857]: _type = "Task" [ 550.189350] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.204031] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.308421] env[61857]: DEBUG nova.policy [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b39f6dff36e46a6aae3afc5c629ad97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75fa9787473f4187a303fc4f036e2aef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 550.447291] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31aa9e85-49a1-4a45-9afe-7358b6eed9e5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.458568] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505a91a2-e858-4861-bf70-406ea846927c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.467686] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Acquiring lock "d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.467686] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Lock "d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.500855] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbea88d2-706e-4403-ba25-d7c0424a7bc2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.515159] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10390150-05de-4662-9554-fc32b5a381e0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.532582] env[61857]: DEBUG nova.compute.provider_tree [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.538485] env[61857]: INFO nova.scheduler.client.report [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Deleted allocations for instance 61faaebe-374a-40ba-aefb-b199ea4dea93 [ 550.669662] env[61857]: DEBUG nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 550.711641] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950677, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.048734] env[61857]: DEBUG nova.scheduler.client.report [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.054404] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1cfd1f25-1425-409c-9d8c-ed3755a32cb1 tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "61faaebe-374a-40ba-aefb-b199ea4dea93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.916s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.209462] env[61857]: DEBUG oslo_vmware.api [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950677, 'name': PowerOnVM_Task, 'duration_secs': 0.567545} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.209805] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 551.209934] env[61857]: DEBUG nova.compute.manager [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 551.210750] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d40f99-b086-4015-89b8-715033f99183 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.294851] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Successfully created port: d54775be-555a-4588-8282-22e67b6b3d2a {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.556816] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.909s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.557760] env[61857]: ERROR nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4b0da35d-33e2-494e-a2da-473d4e03a462, please check neutron logs for more information. [ 551.557760] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Traceback (most recent call last): [ 551.557760] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.557760] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] self.driver.spawn(context, instance, image_meta, [ 551.557760] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 551.557760] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.557760] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.557760] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] vm_ref = self.build_virtual_machine(instance, [ 551.557760] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.557760] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.557760] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] for vif in network_info: [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] return self._sync_wrapper(fn, *args, **kwargs) [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] self.wait() [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] self[:] = self._gt.wait() [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] return self._exit_event.wait() [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] result = hub.switch() [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.559044] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] return self.greenlet.switch() [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] result = function(*args, **kwargs) [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] return func(*args, **kwargs) [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] raise e [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] nwinfo = self.network_api.allocate_for_instance( [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] created_port_ids = self._update_ports_for_instance( [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] with excutils.save_and_reraise_exception(): [ 551.561527] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] self.force_reraise() [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] raise self.value [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] updated_port = self._update_port( [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] _ensure_no_port_binding_failure(port) [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] raise exception.PortBindingFailed(port_id=port['id']) [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] nova.exception.PortBindingFailed: Binding failed for port 4b0da35d-33e2-494e-a2da-473d4e03a462, please check neutron logs for more information. [ 551.562117] env[61857]: ERROR nova.compute.manager [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] [ 551.562921] env[61857]: DEBUG nova.compute.utils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Binding failed for port 4b0da35d-33e2-494e-a2da-473d4e03a462, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 551.563550] env[61857]: DEBUG nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Build of instance 90b0c4ad-902e-425d-b66b-51918e6476d9 was re-scheduled: Binding failed for port 4b0da35d-33e2-494e-a2da-473d4e03a462, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 551.564994] env[61857]: DEBUG nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 551.568424] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquiring lock "refresh_cache-90b0c4ad-902e-425d-b66b-51918e6476d9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.568424] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquired lock "refresh_cache-90b0c4ad-902e-425d-b66b-51918e6476d9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.568424] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 551.568424] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.114s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.570928] env[61857]: INFO nova.compute.claims [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.574768] env[61857]: DEBUG nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 551.687362] env[61857]: DEBUG nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 551.718018] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 551.718018] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 551.718018] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.718219] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 551.718219] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.718219] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 551.718615] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 551.718924] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 551.719225] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 551.719496] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 551.722016] env[61857]: DEBUG nova.virt.hardware [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 551.722016] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f36923-5d8d-48d0-9f29-3bf181eeb5bd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.737919] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83886385-a6da-4cff-ac70-319813814a16 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.745610] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.117528] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.130834] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.228490] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "9f079e2f-a22d-48b7-926a-b89ba087d45d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.228772] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "9f079e2f-a22d-48b7-926a-b89ba087d45d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.232162] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.740353] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Releasing lock "refresh_cache-90b0c4ad-902e-425d-b66b-51918e6476d9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.740561] env[61857]: DEBUG nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 552.743167] env[61857]: DEBUG nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.743167] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 552.769726] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.897141] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fee2be2-7b11-457e-a01d-8206b80783ab {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.909743] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc11ac6d-15c2-4000-8e07-34cdc515962c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.956742] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0716f7-1164-4cf3-9425-ad5f9c27daac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.962818] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Acquiring lock "0ae90dae-d095-418c-abda-f6bb76f89919" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.963070] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Lock "0ae90dae-d095-418c-abda-f6bb76f89919" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.969779] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232586a1-1529-408c-8a79-a4e7d56308f9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.985877] env[61857]: DEBUG nova.compute.provider_tree [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.271788] env[61857]: DEBUG nova.network.neutron [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.492809] env[61857]: DEBUG nova.scheduler.client.report [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.777623] env[61857]: INFO nova.compute.manager [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: 90b0c4ad-902e-425d-b66b-51918e6476d9] Took 1.03 seconds to deallocate network for instance. [ 554.000813] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.000813] env[61857]: DEBUG nova.compute.manager [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 554.001801] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.732s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.311092] env[61857]: ERROR nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d54775be-555a-4588-8282-22e67b6b3d2a, please check neutron logs for more information. [ 554.311092] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 554.311092] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.311092] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 554.311092] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.311092] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 554.311092] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.311092] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 554.311092] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.311092] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 554.311092] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.311092] env[61857]: ERROR nova.compute.manager raise self.value [ 554.311092] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.311092] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 554.311092] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.311092] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 554.312118] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.312118] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 554.312118] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d54775be-555a-4588-8282-22e67b6b3d2a, please check neutron logs for more information. [ 554.312118] env[61857]: ERROR nova.compute.manager [ 554.312118] env[61857]: Traceback (most recent call last): [ 554.312118] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 554.312118] env[61857]: listener.cb(fileno) [ 554.312118] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.312118] env[61857]: result = function(*args, **kwargs) [ 554.312118] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.312118] env[61857]: return func(*args, **kwargs) [ 554.312118] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.312118] env[61857]: raise e [ 554.312118] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.312118] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 554.312118] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.312118] env[61857]: created_port_ids = self._update_ports_for_instance( [ 554.312118] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.312118] env[61857]: with excutils.save_and_reraise_exception(): [ 554.312118] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.312118] env[61857]: self.force_reraise() [ 554.312118] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.312118] env[61857]: raise self.value [ 554.312118] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.312118] env[61857]: updated_port = self._update_port( [ 554.312118] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.312118] env[61857]: _ensure_no_port_binding_failure(port) [ 554.312118] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.312118] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 554.313642] env[61857]: nova.exception.PortBindingFailed: Binding failed for port d54775be-555a-4588-8282-22e67b6b3d2a, please check neutron logs for more information. [ 554.313642] env[61857]: Removing descriptor: 17 [ 554.313642] env[61857]: ERROR nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d54775be-555a-4588-8282-22e67b6b3d2a, please check neutron logs for more information. [ 554.313642] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Traceback (most recent call last): [ 554.313642] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 554.313642] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] yield resources [ 554.313642] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 554.313642] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] self.driver.spawn(context, instance, image_meta, [ 554.313642] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 554.313642] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.313642] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.313642] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] vm_ref = self.build_virtual_machine(instance, [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] for vif in network_info: [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] return self._sync_wrapper(fn, *args, **kwargs) [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] self.wait() [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] self[:] = self._gt.wait() [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] return self._exit_event.wait() [ 554.314230] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] result = hub.switch() [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] return self.greenlet.switch() [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] result = function(*args, **kwargs) [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] return func(*args, **kwargs) [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] raise e [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] nwinfo = self.network_api.allocate_for_instance( [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.314871] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] created_port_ids = self._update_ports_for_instance( [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] with excutils.save_and_reraise_exception(): [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] self.force_reraise() [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] raise self.value [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] updated_port = self._update_port( [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] _ensure_no_port_binding_failure(port) [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.315325] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] raise exception.PortBindingFailed(port_id=port['id']) [ 554.315784] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] nova.exception.PortBindingFailed: Binding failed for port d54775be-555a-4588-8282-22e67b6b3d2a, please check neutron logs for more information. [ 554.315784] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] [ 554.315784] env[61857]: INFO nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Terminating instance [ 554.315784] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Acquiring lock "refresh_cache-ad56d32b-c78d-45db-872a-1c4ea91ab909" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.315784] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Acquired lock "refresh_cache-ad56d32b-c78d-45db-872a-1c4ea91ab909" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.315784] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 554.510904] env[61857]: DEBUG nova.compute.utils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 554.516106] env[61857]: DEBUG nova.compute.manager [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Not allocating networking since 'none' was specified. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 554.561102] env[61857]: INFO nova.compute.manager [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Rebuilding instance [ 554.705444] env[61857]: DEBUG nova.compute.manager [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 554.709222] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d37c49-e847-4452-9f1f-a702fdaf2f38 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.825017] env[61857]: INFO nova.scheduler.client.report [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Deleted allocations for instance 90b0c4ad-902e-425d-b66b-51918e6476d9 [ 554.870184] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.909543] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a956ffc-10d6-4ffd-a4b6-4626e613141a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.916979] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2e2a7b-acf5-4710-8b89-0f493b01cd4e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.950864] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ead7071-0e0f-470a-b052-6627a65ebc79 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.959693] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f80d252-cd8e-4c51-82a9-102023529ce3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.973129] env[61857]: DEBUG nova.compute.provider_tree [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.021104] env[61857]: DEBUG nova.compute.manager [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 555.119907] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Acquiring lock "1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.120175] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Lock "1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.208365] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.225962] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "f14ce7b5-158d-4831-9ba3-404b44752afa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.225962] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "f14ce7b5-158d-4831-9ba3-404b44752afa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.231594] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 555.231838] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c51cc307-2213-4ebf-97f0-e3448b60c439 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.240615] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for the task: (returnval){ [ 555.240615] env[61857]: value = "task-950678" [ 555.240615] env[61857]: _type = "Task" [ 555.240615] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.251276] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950678, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.341277] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9f22bfa7-fc15-4441-ae80-140d476c59a2 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "90b0c4ad-902e-425d-b66b-51918e6476d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.900s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.479196] env[61857]: DEBUG nova.scheduler.client.report [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.558745] env[61857]: DEBUG nova.compute.manager [req-1d59855a-4c94-4a1e-80e1-cdbbecf0c757 req-89d02aa5-f8f6-44e9-be29-1e17244f4807 service nova] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Received event network-changed-d54775be-555a-4588-8282-22e67b6b3d2a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 555.558818] env[61857]: DEBUG nova.compute.manager [req-1d59855a-4c94-4a1e-80e1-cdbbecf0c757 req-89d02aa5-f8f6-44e9-be29-1e17244f4807 service nova] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Refreshing instance network info cache due to event network-changed-d54775be-555a-4588-8282-22e67b6b3d2a. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 555.559082] env[61857]: DEBUG oslo_concurrency.lockutils [req-1d59855a-4c94-4a1e-80e1-cdbbecf0c757 req-89d02aa5-f8f6-44e9-be29-1e17244f4807 service nova] Acquiring lock "refresh_cache-ad56d32b-c78d-45db-872a-1c4ea91ab909" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.710976] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Releasing lock "refresh_cache-ad56d32b-c78d-45db-872a-1c4ea91ab909" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.711454] env[61857]: DEBUG nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 555.711649] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 555.711964] env[61857]: DEBUG oslo_concurrency.lockutils [req-1d59855a-4c94-4a1e-80e1-cdbbecf0c757 req-89d02aa5-f8f6-44e9-be29-1e17244f4807 service nova] Acquired lock "refresh_cache-ad56d32b-c78d-45db-872a-1c4ea91ab909" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.712158] env[61857]: DEBUG nova.network.neutron [req-1d59855a-4c94-4a1e-80e1-cdbbecf0c757 req-89d02aa5-f8f6-44e9-be29-1e17244f4807 service nova] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Refreshing network info cache for port d54775be-555a-4588-8282-22e67b6b3d2a {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 555.713619] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a30877e-ebd6-457f-a9a7-e5f190b8e0e5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.726026] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4c028e-e173-43f1-8ce3-da523d795cb9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.754315] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad56d32b-c78d-45db-872a-1c4ea91ab909 could not be found. [ 555.754577] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 555.754778] env[61857]: INFO nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Took 0.04 seconds to destroy the instance on the hypervisor. [ 555.755079] env[61857]: DEBUG oslo.service.loopingcall [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.758739] env[61857]: DEBUG nova.compute.manager [-] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 555.758739] env[61857]: DEBUG nova.network.neutron [-] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 555.764561] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950678, 'name': PowerOffVM_Task, 'duration_secs': 0.126364} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.765121] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 555.765371] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 555.766518] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d98ecf-aace-4ff5-aaf3-228b530fd9c5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.773993] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 555.777011] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ae87f07-fe3d-4ee6-8b60-20da163cb8ad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.798311] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 555.798311] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 555.798311] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Deleting the datastore file [datastore1] bb418d62-d47c-42c5-908e-26f1eb034e46 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 555.798311] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05f37dfc-0f58-41aa-81bf-a198ac2521d5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.806969] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for the task: (returnval){ [ 555.806969] env[61857]: value = "task-950680" [ 555.806969] env[61857]: _type = "Task" [ 555.806969] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.811640] env[61857]: DEBUG nova.network.neutron [-] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.818965] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950680, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.844254] env[61857]: DEBUG nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 555.986835] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.985s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.987776] env[61857]: ERROR nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04, please check neutron logs for more information. [ 555.987776] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Traceback (most recent call last): [ 555.987776] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 555.987776] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] self.driver.spawn(context, instance, image_meta, [ 555.987776] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 555.987776] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.987776] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.987776] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] vm_ref = self.build_virtual_machine(instance, [ 555.987776] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.987776] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.987776] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] for vif in network_info: [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] return self._sync_wrapper(fn, *args, **kwargs) [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] self.wait() [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] self[:] = self._gt.wait() [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] return self._exit_event.wait() [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] result = hub.switch() [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.988116] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] return self.greenlet.switch() [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] result = function(*args, **kwargs) [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] return func(*args, **kwargs) [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] raise e [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] nwinfo = self.network_api.allocate_for_instance( [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] created_port_ids = self._update_ports_for_instance( [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] with excutils.save_and_reraise_exception(): [ 555.988482] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] self.force_reraise() [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] raise self.value [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] updated_port = self._update_port( [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] _ensure_no_port_binding_failure(port) [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] raise exception.PortBindingFailed(port_id=port['id']) [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] nova.exception.PortBindingFailed: Binding failed for port 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04, please check neutron logs for more information. [ 555.988824] env[61857]: ERROR nova.compute.manager [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] [ 555.989289] env[61857]: DEBUG nova.compute.utils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Binding failed for port 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 555.990673] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.096s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.994356] env[61857]: DEBUG nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Build of instance b34ca3cc-d26d-4068-a27d-a6d66404dfcb was re-scheduled: Binding failed for port 808e5c82-8f76-4fda-b28c-9aaa6a1a5f04, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 555.994902] env[61857]: DEBUG nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 555.995133] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Acquiring lock "refresh_cache-b34ca3cc-d26d-4068-a27d-a6d66404dfcb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.995281] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Acquired lock "refresh_cache-b34ca3cc-d26d-4068-a27d-a6d66404dfcb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.995437] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 556.041380] env[61857]: DEBUG nova.compute.manager [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 556.085169] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.085311] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.085483] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.085874] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.085874] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.085962] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.086648] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.086648] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.086648] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.086648] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.086841] env[61857]: DEBUG nova.virt.hardware [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.088135] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd56af52-b35d-432c-b4de-bea45ac25b51 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.097143] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc091035-2de0-4938-a12a-09fbbf6b78c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.121222] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 556.126758] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Creating folder: Project (ac441e844ea547429088a011c8269bfc). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 556.127021] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ebc96af-5eb7-45f5-a73a-25f7157851d1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.139028] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Created folder: Project (ac441e844ea547429088a011c8269bfc) in parent group-v214027. [ 556.139028] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Creating folder: Instances. Parent ref: group-v214032. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 556.139028] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf4d6586-5048-4443-8c4b-ed7076639a9f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.150715] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Created folder: Instances in parent group-v214032. [ 556.150715] env[61857]: DEBUG oslo.service.loopingcall [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 556.150715] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 556.150715] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fec5d98a-62c8-43fd-8893-9ca8371d9145 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.171353] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 556.171353] env[61857]: value = "task-950683" [ 556.171353] env[61857]: _type = "Task" [ 556.171353] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.178848] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950683, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.263501] env[61857]: DEBUG nova.network.neutron [req-1d59855a-4c94-4a1e-80e1-cdbbecf0c757 req-89d02aa5-f8f6-44e9-be29-1e17244f4807 service nova] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.316674] env[61857]: DEBUG nova.network.neutron [-] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.323942] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950680, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254574} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.324236] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 556.324406] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 556.324759] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 556.373711] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.419150] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Acquiring lock "5ad7fc93-72b3-4108-923d-f52d74589423" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.419403] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Lock "5ad7fc93-72b3-4108-923d-f52d74589423" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.468882] env[61857]: DEBUG nova.network.neutron [req-1d59855a-4c94-4a1e-80e1-cdbbecf0c757 req-89d02aa5-f8f6-44e9-be29-1e17244f4807 service nova] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.526046] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.648421] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.688013] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950683, 'name': CreateVM_Task, 'duration_secs': 0.289134} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.688904] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 556.689742] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.689896] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.690265] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 556.690506] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f216ba64-7620-4780-8f16-23556b1ccc0f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.695260] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for the task: (returnval){ [ 556.695260] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5246a273-8752-88a5-a426-549fb749cb5f" [ 556.695260] env[61857]: _type = "Task" [ 556.695260] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.706230] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5246a273-8752-88a5-a426-549fb749cb5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.819322] env[61857]: INFO nova.compute.manager [-] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Took 1.06 seconds to deallocate network for instance. [ 556.820172] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60298d10-5064-4be3-825c-a978a927565d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.830092] env[61857]: DEBUG nova.compute.claims [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 556.830092] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.837593] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746fd82e-3891-413b-a063-5abb691513e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.884760] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cb4d77-2ba9-49d9-aed9-4bcbc80525fa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.894726] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958ecce5-62bc-4e21-a541-c2f4654a3c82 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.913447] env[61857]: DEBUG nova.compute.provider_tree [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.974545] env[61857]: DEBUG oslo_concurrency.lockutils [req-1d59855a-4c94-4a1e-80e1-cdbbecf0c757 req-89d02aa5-f8f6-44e9-be29-1e17244f4807 service nova] Releasing lock "refresh_cache-ad56d32b-c78d-45db-872a-1c4ea91ab909" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.150902] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Releasing lock "refresh_cache-b34ca3cc-d26d-4068-a27d-a6d66404dfcb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.151348] env[61857]: DEBUG nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 557.151348] env[61857]: DEBUG nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.151499] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 557.173282] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.209643] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5246a273-8752-88a5-a426-549fb749cb5f, 'name': SearchDatastore_Task, 'duration_secs': 0.008911} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.210024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.210291] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 557.210524] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.210660] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.210834] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 557.211128] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2781c36-9829-4809-b5dd-0325d85ec6fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.219458] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 557.220193] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 557.220723] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b2ba523-07d6-4bda-b9d4-dc6ed90d8203 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.228191] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "b54af0ef-448e-453c-9056-32ad9141d9d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.228429] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "b54af0ef-448e-453c-9056-32ad9141d9d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.233901] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for the task: (returnval){ [ 557.233901] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52da903d-8483-3f48-b08d-37f46a6bca3f" [ 557.233901] env[61857]: _type = "Task" [ 557.233901] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.242110] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52da903d-8483-3f48-b08d-37f46a6bca3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.367132] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 557.367410] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 557.368464] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.369075] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 557.369503] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.369844] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 557.370156] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 557.370468] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 557.370707] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 557.371063] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 557.371267] env[61857]: DEBUG nova.virt.hardware [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 557.372138] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0b10fd-e4b2-496a-8468-026998e853e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.387923] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a556db9-8ec6-4118-979f-b36e9744519e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.402426] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 557.410870] env[61857]: DEBUG oslo.service.loopingcall [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.411220] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 557.412846] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a4098691-9d69-4164-bb69-626eccd40e6b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.427023] env[61857]: DEBUG nova.scheduler.client.report [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.436227] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 557.436227] env[61857]: value = "task-950684" [ 557.436227] env[61857]: _type = "Task" [ 557.436227] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.444274] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950684, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.552505] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Acquiring lock "c1526323-a057-4a84-aa32-05cd802ebf9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.552797] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Lock "c1526323-a057-4a84-aa32-05cd802ebf9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.678307] env[61857]: DEBUG nova.network.neutron [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.747589] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52da903d-8483-3f48-b08d-37f46a6bca3f, 'name': SearchDatastore_Task, 'duration_secs': 0.02114} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.748235] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ce1ff5a-ee43-4fb7-b600-081b34b29f50 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.756308] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for the task: (returnval){ [ 557.756308] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fad91f-da54-68f5-3a6a-7adb99053b69" [ 557.756308] env[61857]: _type = "Task" [ 557.756308] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.768256] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fad91f-da54-68f5-3a6a-7adb99053b69, 'name': SearchDatastore_Task, 'duration_secs': 0.008297} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.768552] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.768821] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 2e7c154b-dab2-4384-bdb8-d098967654c5/2e7c154b-dab2-4384-bdb8-d098967654c5.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 557.769110] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d988341-3e04-4b71-ada7-a267fa8143c5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.776546] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for the task: (returnval){ [ 557.776546] env[61857]: value = "task-950685" [ 557.776546] env[61857]: _type = "Task" [ 557.776546] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.785167] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.932469] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.942s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.933485] env[61857]: ERROR nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bba41198-8b1a-4895-ae13-fd458fd4eb5d, please check neutron logs for more information. [ 557.933485] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Traceback (most recent call last): [ 557.933485] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.933485] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] self.driver.spawn(context, instance, image_meta, [ 557.933485] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 557.933485] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.933485] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.933485] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] vm_ref = self.build_virtual_machine(instance, [ 557.933485] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.933485] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.933485] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] for vif in network_info: [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] return self._sync_wrapper(fn, *args, **kwargs) [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] self.wait() [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] self[:] = self._gt.wait() [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] return self._exit_event.wait() [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] result = hub.switch() [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.934217] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] return self.greenlet.switch() [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] result = function(*args, **kwargs) [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] return func(*args, **kwargs) [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] raise e [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] nwinfo = self.network_api.allocate_for_instance( [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] created_port_ids = self._update_ports_for_instance( [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] with excutils.save_and_reraise_exception(): [ 557.934967] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] self.force_reraise() [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] raise self.value [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] updated_port = self._update_port( [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] _ensure_no_port_binding_failure(port) [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] raise exception.PortBindingFailed(port_id=port['id']) [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] nova.exception.PortBindingFailed: Binding failed for port bba41198-8b1a-4895-ae13-fd458fd4eb5d, please check neutron logs for more information. [ 557.935331] env[61857]: ERROR nova.compute.manager [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] [ 557.935943] env[61857]: DEBUG nova.compute.utils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Binding failed for port bba41198-8b1a-4895-ae13-fd458fd4eb5d, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 557.935943] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.907s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.937729] env[61857]: INFO nova.compute.claims [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 557.946022] env[61857]: DEBUG nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Build of instance b0fd68df-fcba-494d-bf79-a0cf2ea82ed5 was re-scheduled: Binding failed for port bba41198-8b1a-4895-ae13-fd458fd4eb5d, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 557.946022] env[61857]: DEBUG nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 557.946022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.946022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.946576] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 557.958082] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950684, 'name': CreateVM_Task, 'duration_secs': 0.288696} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.958863] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 557.959332] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.959501] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.960364] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 557.960900] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97a4362f-f753-435b-8ecc-568a90f15c0d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.968024] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for the task: (returnval){ [ 557.968024] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521e99fb-a029-0625-77ad-b88d6c5d6cfe" [ 557.968024] env[61857]: _type = "Task" [ 557.968024] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.979591] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521e99fb-a029-0625-77ad-b88d6c5d6cfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.110771] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquiring lock "dafceac5-1eea-4103-8a48-ca6c5136390e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.111411] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "dafceac5-1eea-4103-8a48-ca6c5136390e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.184235] env[61857]: INFO nova.compute.manager [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] [instance: b34ca3cc-d26d-4068-a27d-a6d66404dfcb] Took 1.03 seconds to deallocate network for instance. [ 558.286716] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950685, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441053} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.287694] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 2e7c154b-dab2-4384-bdb8-d098967654c5/2e7c154b-dab2-4384-bdb8-d098967654c5.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 558.287694] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 558.287694] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df027a54-4049-4aa7-8c7a-f31919ef9062 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.294340] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for the task: (returnval){ [ 558.294340] env[61857]: value = "task-950686" [ 558.294340] env[61857]: _type = "Task" [ 558.294340] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.301742] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950686, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.473974] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.483963] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521e99fb-a029-0625-77ad-b88d6c5d6cfe, 'name': SearchDatastore_Task, 'duration_secs': 0.051456} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.483963] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.483963] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 558.483963] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.484117] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.484117] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 558.484117] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c162d2d-c58e-4d1c-a4b6-2a88ffd8f38c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.491352] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 558.491717] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 558.492593] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae6e44e2-96e9-4f24-b511-39c09cd50df5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.499948] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for the task: (returnval){ [ 558.499948] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]526b6c8e-7643-75c0-229c-413e17ea71ff" [ 558.499948] env[61857]: _type = "Task" [ 558.499948] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.509699] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]526b6c8e-7643-75c0-229c-413e17ea71ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.552966] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.703595] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "406f6726-4009-49bd-a37b-91707ece1932" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.704033] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "406f6726-4009-49bd-a37b-91707ece1932" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.807351] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950686, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062646} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.807967] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 558.809310] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448c9a99-c13a-4a1e-b5b9-77eb8c6a3653 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.835305] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] 2e7c154b-dab2-4384-bdb8-d098967654c5/2e7c154b-dab2-4384-bdb8-d098967654c5.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 558.836010] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13db6c1b-45ec-4b27-a1e9-d04164e5efec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.857824] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for the task: (returnval){ [ 558.857824] env[61857]: value = "task-950687" [ 558.857824] env[61857]: _type = "Task" [ 558.857824] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.865886] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950687, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.014691] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]526b6c8e-7643-75c0-229c-413e17ea71ff, 'name': SearchDatastore_Task, 'duration_secs': 0.008456} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.023302] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5979807-f3e3-4567-8205-8dd715b66ee9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.031098] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for the task: (returnval){ [ 559.031098] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52c06784-0de1-187f-1eec-0ee5cbed9b39" [ 559.031098] env[61857]: _type = "Task" [ 559.031098] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.043087] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c06784-0de1-187f-1eec-0ee5cbed9b39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.056769] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.059154] env[61857]: DEBUG nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 559.059154] env[61857]: DEBUG nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.059154] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 559.094823] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.202628] env[61857]: DEBUG nova.compute.manager [req-8c4b095d-a24a-4cd3-8e3f-a176f23f3926 req-8eb8b70e-6bd9-4959-9033-5a234c87477e service nova] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Received event network-vif-deleted-d54775be-555a-4588-8282-22e67b6b3d2a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 559.223352] env[61857]: INFO nova.scheduler.client.report [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Deleted allocations for instance b34ca3cc-d26d-4068-a27d-a6d66404dfcb [ 559.346525] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-321657a2-76fd-40d2-8140-e88eed2f40f9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.354256] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf35765-5b65-467c-85de-366b035e999c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.368256] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950687, 'name': ReconfigVM_Task, 'duration_secs': 0.268317} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.401495] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Reconfigured VM instance instance-0000000a to attach disk [datastore2] 2e7c154b-dab2-4384-bdb8-d098967654c5/2e7c154b-dab2-4384-bdb8-d098967654c5.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 559.401495] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ee42635-634a-488e-b61c-516937127377 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.405274] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ba5a00-8aca-420d-9171-cde1d8449c73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.414461] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5553ee-2edb-4fa5-980c-079262dc2d86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.419452] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for the task: (returnval){ [ 559.419452] env[61857]: value = "task-950688" [ 559.419452] env[61857]: _type = "Task" [ 559.419452] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.431388] env[61857]: DEBUG nova.compute.provider_tree [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.437867] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950688, 'name': Rename_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.517457] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Acquiring lock "3375702e-36fd-43d4-95da-e3aafb96a32f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.517721] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Lock "3375702e-36fd-43d4-95da-e3aafb96a32f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.541470] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c06784-0de1-187f-1eec-0ee5cbed9b39, 'name': SearchDatastore_Task, 'duration_secs': 0.008995} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.542170] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.542170] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 559.542398] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c52a77e5-3fdb-4ac6-8468-e73cf0fd0655 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.548346] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for the task: (returnval){ [ 559.548346] env[61857]: value = "task-950689" [ 559.548346] env[61857]: _type = "Task" [ 559.548346] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.556984] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950689, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.599397] env[61857]: DEBUG nova.network.neutron [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.732223] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e96a20a-528f-4ca2-87f3-8be1a04c6ac0 tempest-ServerDiagnosticsNegativeTest-1045567674 tempest-ServerDiagnosticsNegativeTest-1045567674-project-member] Lock "b34ca3cc-d26d-4068-a27d-a6d66404dfcb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.369s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.938325] env[61857]: DEBUG nova.scheduler.client.report [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.943384] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950688, 'name': Rename_Task, 'duration_secs': 0.139038} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.943938] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 559.944612] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc2ce9f7-ab0a-431b-af92-7c2afe21114f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.958456] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for the task: (returnval){ [ 559.958456] env[61857]: value = "task-950690" [ 559.958456] env[61857]: _type = "Task" [ 559.958456] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.973424] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950690, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.029088] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Acquiring lock "1492bf97-9e03-40eb-ae6c-041f2998124e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.029326] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Lock "1492bf97-9e03-40eb-ae6c-041f2998124e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.058716] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950689, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465978} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.058716] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 560.058716] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 560.058716] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1372dfbf-e7b0-46a9-aaf8-8a6f9dede914 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.066209] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for the task: (returnval){ [ 560.066209] env[61857]: value = "task-950691" [ 560.066209] env[61857]: _type = "Task" [ 560.066209] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.076685] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950691, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.102213] env[61857]: INFO nova.compute.manager [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: b0fd68df-fcba-494d-bf79-a0cf2ea82ed5] Took 1.04 seconds to deallocate network for instance. [ 560.236260] env[61857]: DEBUG nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.445953] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.446536] env[61857]: DEBUG nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 560.449589] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.264s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.469308] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950690, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.579080] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950691, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061456} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.579080] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 560.580360] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91622621-6584-430c-b31b-8b9adba53cf9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.604261] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 560.604261] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f7d9b9e-d9fb-40a9-bb76-8cfa173bf92b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.624655] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for the task: (returnval){ [ 560.624655] env[61857]: value = "task-950692" [ 560.624655] env[61857]: _type = "Task" [ 560.624655] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.634194] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950692, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.771220] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.957926] env[61857]: DEBUG nova.compute.utils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 560.959833] env[61857]: DEBUG nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 560.960030] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 560.974434] env[61857]: DEBUG oslo_vmware.api [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950690, 'name': PowerOnVM_Task, 'duration_secs': 0.673771} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.974623] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 560.974834] env[61857]: INFO nova.compute.manager [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Took 4.94 seconds to spawn the instance on the hypervisor. [ 560.975020] env[61857]: DEBUG nova.compute.manager [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 560.975848] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3875bcee-85b7-45ed-951d-fbb3e2433285 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.053712] env[61857]: DEBUG nova.policy [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f5af5f69b9cb428d9453bc023e16aeaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '18397efabee24a42b6eb6021ab7a04b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 561.138888] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950692, 'name': ReconfigVM_Task, 'duration_secs': 0.272995} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.138888] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Reconfigured VM instance instance-00000001 to attach disk [datastore2] bb418d62-d47c-42c5-908e-26f1eb034e46/bb418d62-d47c-42c5-908e-26f1eb034e46.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 561.138888] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-192298fc-d82b-4055-b2d8-b4eff5e40390 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.150133] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for the task: (returnval){ [ 561.150133] env[61857]: value = "task-950693" [ 561.150133] env[61857]: _type = "Task" [ 561.150133] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.154754] env[61857]: INFO nova.scheduler.client.report [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleted allocations for instance b0fd68df-fcba-494d-bf79-a0cf2ea82ed5 [ 561.175151] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950693, 'name': Rename_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.374019] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d175b9-4566-470d-abcf-f3e4787035af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.383117] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df953d61-52b5-48a9-8238-3b3bcdfbdb17 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.417882] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f741b37-d0ca-47c8-a416-dc974067b336 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.424888] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616d18a0-3706-4989-9d16-b50d2b3be908 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.431109] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Successfully created port: 0a3cb614-e979-4b79-86df-dbf4a64939ab {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 561.441961] env[61857]: DEBUG nova.compute.provider_tree [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.460250] env[61857]: DEBUG nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 561.504866] env[61857]: INFO nova.compute.manager [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Took 19.08 seconds to build instance. [ 561.665172] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950693, 'name': Rename_Task, 'duration_secs': 0.144072} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.665607] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 561.665734] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a24d148c-a1a1-48a0-90bc-30710ee9cb4d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.669107] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ac8b75fe-0f83-4d1b-b6a1-788c686df383 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "b0fd68df-fcba-494d-bf79-a0cf2ea82ed5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.398s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.673082] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Waiting for the task: (returnval){ [ 561.673082] env[61857]: value = "task-950694" [ 561.673082] env[61857]: _type = "Task" [ 561.673082] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.684071] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950694, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.951157] env[61857]: DEBUG nova.scheduler.client.report [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.008590] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8f09285e-3b96-451d-af6f-7fdeca44abeb tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lock "2e7c154b-dab2-4384-bdb8-d098967654c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.594s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.173039] env[61857]: DEBUG nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.197746] env[61857]: DEBUG oslo_vmware.api [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Task: {'id': task-950694, 'name': PowerOnVM_Task, 'duration_secs': 0.470775} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.197746] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 562.197746] env[61857]: DEBUG nova.compute.manager [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 562.198533] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8297d2-405c-4e3b-b84e-b92a31e9796c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.463266] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.463488] env[61857]: ERROR nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 27d6f88e-1abe-4552-b7a1-839802105483, please check neutron logs for more information. [ 562.463488] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Traceback (most recent call last): [ 562.463488] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 562.463488] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] self.driver.spawn(context, instance, image_meta, [ 562.463488] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 562.463488] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.463488] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.463488] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] vm_ref = self.build_virtual_machine(instance, [ 562.463488] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.463488] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.463488] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] for vif in network_info: [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] return self._sync_wrapper(fn, *args, **kwargs) [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] self.wait() [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] self[:] = self._gt.wait() [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] return self._exit_event.wait() [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] result = hub.switch() [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.463939] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] return self.greenlet.switch() [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] result = function(*args, **kwargs) [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] return func(*args, **kwargs) [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] raise e [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] nwinfo = self.network_api.allocate_for_instance( [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] created_port_ids = self._update_ports_for_instance( [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] with excutils.save_and_reraise_exception(): [ 562.465285] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] self.force_reraise() [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] raise self.value [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] updated_port = self._update_port( [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] _ensure_no_port_binding_failure(port) [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] raise exception.PortBindingFailed(port_id=port['id']) [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] nova.exception.PortBindingFailed: Binding failed for port 27d6f88e-1abe-4552-b7a1-839802105483, please check neutron logs for more information. [ 562.465804] env[61857]: ERROR nova.compute.manager [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] [ 562.466287] env[61857]: DEBUG nova.compute.utils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Binding failed for port 27d6f88e-1abe-4552-b7a1-839802105483, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 562.466287] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.236s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.468281] env[61857]: DEBUG nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Build of instance 6acac8fe-a2a7-426e-912a-0292d351887e was re-scheduled: Binding failed for port 27d6f88e-1abe-4552-b7a1-839802105483, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 562.468720] env[61857]: DEBUG nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 562.469218] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Acquiring lock "refresh_cache-6acac8fe-a2a7-426e-912a-0292d351887e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.469401] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Acquired lock "refresh_cache-6acac8fe-a2a7-426e-912a-0292d351887e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.469567] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 562.476429] env[61857]: DEBUG nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 562.506989] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 562.507259] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 562.507421] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.507625] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 562.508070] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.508070] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 562.508167] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 562.508348] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 562.508930] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 562.508930] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 562.508930] env[61857]: DEBUG nova.virt.hardware [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 562.510215] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1383c221-40ba-4538-b238-b24abce86c75 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.514788] env[61857]: DEBUG nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.522947] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddabb252-5fe9-45b2-aea1-32f4382bac65 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.706453] env[61857]: ERROR nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0a3cb614-e979-4b79-86df-dbf4a64939ab, please check neutron logs for more information. [ 562.706453] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 562.706453] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.706453] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 562.706453] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 562.706453] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 562.706453] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 562.706453] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 562.706453] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.706453] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 562.706453] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.706453] env[61857]: ERROR nova.compute.manager raise self.value [ 562.706453] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 562.706453] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 562.706453] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.706453] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 562.707430] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.707430] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 562.707430] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0a3cb614-e979-4b79-86df-dbf4a64939ab, please check neutron logs for more information. [ 562.707430] env[61857]: ERROR nova.compute.manager [ 562.707430] env[61857]: Traceback (most recent call last): [ 562.707430] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 562.707430] env[61857]: listener.cb(fileno) [ 562.707430] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.707430] env[61857]: result = function(*args, **kwargs) [ 562.707430] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.707430] env[61857]: return func(*args, **kwargs) [ 562.707430] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.707430] env[61857]: raise e [ 562.707430] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.707430] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 562.707430] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 562.707430] env[61857]: created_port_ids = self._update_ports_for_instance( [ 562.707430] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 562.707430] env[61857]: with excutils.save_and_reraise_exception(): [ 562.707430] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.707430] env[61857]: self.force_reraise() [ 562.707430] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.707430] env[61857]: raise self.value [ 562.707430] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 562.707430] env[61857]: updated_port = self._update_port( [ 562.707430] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.707430] env[61857]: _ensure_no_port_binding_failure(port) [ 562.707430] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.707430] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 562.708743] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 0a3cb614-e979-4b79-86df-dbf4a64939ab, please check neutron logs for more information. [ 562.708743] env[61857]: Removing descriptor: 17 [ 562.708743] env[61857]: ERROR nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0a3cb614-e979-4b79-86df-dbf4a64939ab, please check neutron logs for more information. [ 562.708743] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Traceback (most recent call last): [ 562.708743] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 562.708743] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] yield resources [ 562.708743] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 562.708743] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] self.driver.spawn(context, instance, image_meta, [ 562.708743] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 562.708743] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.708743] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.708743] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] vm_ref = self.build_virtual_machine(instance, [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] for vif in network_info: [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] return self._sync_wrapper(fn, *args, **kwargs) [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] self.wait() [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] self[:] = self._gt.wait() [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] return self._exit_event.wait() [ 562.709211] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] result = hub.switch() [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] return self.greenlet.switch() [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] result = function(*args, **kwargs) [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] return func(*args, **kwargs) [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] raise e [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] nwinfo = self.network_api.allocate_for_instance( [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 562.709553] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] created_port_ids = self._update_ports_for_instance( [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] with excutils.save_and_reraise_exception(): [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] self.force_reraise() [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] raise self.value [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] updated_port = self._update_port( [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] _ensure_no_port_binding_failure(port) [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.709868] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] raise exception.PortBindingFailed(port_id=port['id']) [ 562.711259] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] nova.exception.PortBindingFailed: Binding failed for port 0a3cb614-e979-4b79-86df-dbf4a64939ab, please check neutron logs for more information. [ 562.711259] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] [ 562.711259] env[61857]: INFO nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Terminating instance [ 562.715456] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquiring lock "refresh_cache-e8c59da6-c226-4c83-869b-6a200cf1c0b9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.715655] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquired lock "refresh_cache-e8c59da6-c226-4c83-869b-6a200cf1c0b9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.715853] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 562.717588] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.724393] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.865670] env[61857]: DEBUG nova.compute.manager [None req-cb020c4b-fda3-4991-918a-29389e905ec6 tempest-ServerDiagnosticsV248Test-2033239442 tempest-ServerDiagnosticsV248Test-2033239442-project-admin] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 562.869187] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cbacb2c-0636-40ca-ba4a-bb14288b99f1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.878026] env[61857]: INFO nova.compute.manager [None req-cb020c4b-fda3-4991-918a-29389e905ec6 tempest-ServerDiagnosticsV248Test-2033239442 tempest-ServerDiagnosticsV248Test-2033239442-project-admin] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Retrieving diagnostics [ 562.878926] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248ddf1e-27a0-449b-969b-dd891c92671d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.013891] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 563.042730] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.091356] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.245608] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 563.270371] env[61857]: DEBUG nova.compute.manager [req-273af879-6730-4943-8028-65c5fce23cca req-a311abb7-cea3-41b8-a4fb-70fb95508ae4 service nova] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Received event network-changed-0a3cb614-e979-4b79-86df-dbf4a64939ab {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 563.270551] env[61857]: DEBUG nova.compute.manager [req-273af879-6730-4943-8028-65c5fce23cca req-a311abb7-cea3-41b8-a4fb-70fb95508ae4 service nova] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Refreshing instance network info cache due to event network-changed-0a3cb614-e979-4b79-86df-dbf4a64939ab. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 563.270734] env[61857]: DEBUG oslo_concurrency.lockutils [req-273af879-6730-4943-8028-65c5fce23cca req-a311abb7-cea3-41b8-a4fb-70fb95508ae4 service nova] Acquiring lock "refresh_cache-e8c59da6-c226-4c83-869b-6a200cf1c0b9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.327091] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f862cd35-dad7-4e05-833c-1821ce164fff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.338099] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e02f765e-46e3-4d42-ad68-76d23a66aff4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.372180] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.373719] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcc7a36-765b-48e4-aac0-df4841d0104d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.381800] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b493bc-ff17-4abe-be0c-0d3f8d8e53d5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.397576] env[61857]: DEBUG nova.compute.provider_tree [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.596863] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Releasing lock "refresh_cache-6acac8fe-a2a7-426e-912a-0292d351887e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.597210] env[61857]: DEBUG nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 563.597525] env[61857]: DEBUG nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.597763] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 563.614618] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 563.877716] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Releasing lock "refresh_cache-e8c59da6-c226-4c83-869b-6a200cf1c0b9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.880035] env[61857]: DEBUG nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 563.880035] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 563.880508] env[61857]: DEBUG oslo_concurrency.lockutils [req-273af879-6730-4943-8028-65c5fce23cca req-a311abb7-cea3-41b8-a4fb-70fb95508ae4 service nova] Acquired lock "refresh_cache-e8c59da6-c226-4c83-869b-6a200cf1c0b9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.880717] env[61857]: DEBUG nova.network.neutron [req-273af879-6730-4943-8028-65c5fce23cca req-a311abb7-cea3-41b8-a4fb-70fb95508ae4 service nova] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Refreshing network info cache for port 0a3cb614-e979-4b79-86df-dbf4a64939ab {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 563.881851] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91dd02c0-3404-47da-b35a-e34e1729a5cc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.893580] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e2ecb8-3f26-4a04-a4cc-925ca02627fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.910933] env[61857]: DEBUG nova.scheduler.client.report [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.930187] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e8c59da6-c226-4c83-869b-6a200cf1c0b9 could not be found. [ 563.930187] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 563.930944] env[61857]: INFO nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 563.930944] env[61857]: DEBUG oslo.service.loopingcall [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.931680] env[61857]: DEBUG nova.compute.manager [-] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.931680] env[61857]: DEBUG nova.network.neutron [-] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 563.964079] env[61857]: DEBUG nova.network.neutron [-] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.116719] env[61857]: DEBUG nova.network.neutron [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.305152] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquiring lock "ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.305524] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.408801] env[61857]: DEBUG nova.network.neutron [req-273af879-6730-4943-8028-65c5fce23cca req-a311abb7-cea3-41b8-a4fb-70fb95508ae4 service nova] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.418617] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.418617] env[61857]: ERROR nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 443c1a0b-6379-41eb-95ce-81de874f8587, please check neutron logs for more information. [ 564.418617] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Traceback (most recent call last): [ 564.418617] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 564.418617] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] self.driver.spawn(context, instance, image_meta, [ 564.418617] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 564.418617] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 564.418617] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 564.418617] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] vm_ref = self.build_virtual_machine(instance, [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] for vif in network_info: [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] return self._sync_wrapper(fn, *args, **kwargs) [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] self.wait() [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] self[:] = self._gt.wait() [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] return self._exit_event.wait() [ 564.419027] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] result = hub.switch() [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] return self.greenlet.switch() [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] result = function(*args, **kwargs) [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] return func(*args, **kwargs) [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] raise e [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] nwinfo = self.network_api.allocate_for_instance( [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 564.419422] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] created_port_ids = self._update_ports_for_instance( [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] with excutils.save_and_reraise_exception(): [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] self.force_reraise() [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] raise self.value [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] updated_port = self._update_port( [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] _ensure_no_port_binding_failure(port) [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.419747] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] raise exception.PortBindingFailed(port_id=port['id']) [ 564.420121] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] nova.exception.PortBindingFailed: Binding failed for port 443c1a0b-6379-41eb-95ce-81de874f8587, please check neutron logs for more information. [ 564.420121] env[61857]: ERROR nova.compute.manager [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] [ 564.420121] env[61857]: DEBUG nova.compute.utils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Binding failed for port 443c1a0b-6379-41eb-95ce-81de874f8587, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 564.421367] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.188s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.428322] env[61857]: DEBUG nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Build of instance 4c7dab9e-4527-46b6-a1ec-6e6935768b75 was re-scheduled: Binding failed for port 443c1a0b-6379-41eb-95ce-81de874f8587, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 564.428322] env[61857]: DEBUG nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 564.428322] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Acquiring lock "refresh_cache-4c7dab9e-4527-46b6-a1ec-6e6935768b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.428322] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Acquired lock "refresh_cache-4c7dab9e-4527-46b6-a1ec-6e6935768b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.428577] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 564.469028] env[61857]: DEBUG nova.network.neutron [-] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.535541] env[61857]: DEBUG nova.network.neutron [req-273af879-6730-4943-8028-65c5fce23cca req-a311abb7-cea3-41b8-a4fb-70fb95508ae4 service nova] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.624590] env[61857]: INFO nova.compute.manager [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] [instance: 6acac8fe-a2a7-426e-912a-0292d351887e] Took 1.03 seconds to deallocate network for instance. [ 564.979196] env[61857]: INFO nova.compute.manager [-] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Took 1.05 seconds to deallocate network for instance. [ 564.991205] env[61857]: DEBUG nova.compute.claims [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 564.991205] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.991684] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.037412] env[61857]: DEBUG oslo_concurrency.lockutils [req-273af879-6730-4943-8028-65c5fce23cca req-a311abb7-cea3-41b8-a4fb-70fb95508ae4 service nova] Releasing lock "refresh_cache-e8c59da6-c226-4c83-869b-6a200cf1c0b9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.060776] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "7e466be0-f291-43c3-a79c-fb3871f06c23" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.061033] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "7e466be0-f291-43c3-a79c-fb3871f06c23" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.236366] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.411546] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3bfacc0-a1d1-46fe-b3da-cb66722c637f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.420444] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45060a72-478b-437b-87a8-57cfab5d0e24 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.455858] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effc4d14-6c0d-4851-8d4c-ec410cdaada2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.464117] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c014098-2a87-4dc9-b9cc-6a8d332acc2c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.482080] env[61857]: DEBUG nova.compute.provider_tree [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.659461] env[61857]: INFO nova.scheduler.client.report [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Deleted allocations for instance 6acac8fe-a2a7-426e-912a-0292d351887e [ 565.745235] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Releasing lock "refresh_cache-4c7dab9e-4527-46b6-a1ec-6e6935768b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.745235] env[61857]: DEBUG nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 565.745235] env[61857]: DEBUG nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 565.745235] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 565.791790] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.932099] env[61857]: DEBUG nova.compute.manager [req-1ded5bc6-594a-4d40-9164-d7441c06251a req-adf56463-e8f3-4922-b675-4c3077bfce4e service nova] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Received event network-vif-deleted-0a3cb614-e979-4b79-86df-dbf4a64939ab {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 565.986307] env[61857]: DEBUG nova.scheduler.client.report [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.015124] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "bb418d62-d47c-42c5-908e-26f1eb034e46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.015420] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "bb418d62-d47c-42c5-908e-26f1eb034e46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.015636] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "bb418d62-d47c-42c5-908e-26f1eb034e46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.016773] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "bb418d62-d47c-42c5-908e-26f1eb034e46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.016773] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "bb418d62-d47c-42c5-908e-26f1eb034e46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.018412] env[61857]: INFO nova.compute.manager [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Terminating instance [ 566.020517] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "refresh_cache-bb418d62-d47c-42c5-908e-26f1eb034e46" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.020921] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquired lock "refresh_cache-bb418d62-d47c-42c5-908e-26f1eb034e46" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.021178] env[61857]: DEBUG nova.network.neutron [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 566.170421] env[61857]: DEBUG oslo_concurrency.lockutils [None req-54997f4f-1bc7-4819-bed3-ef62a5235cab tempest-ServersAdminNegativeTestJSON-1183736504 tempest-ServersAdminNegativeTestJSON-1183736504-project-member] Lock "6acac8fe-a2a7-426e-912a-0292d351887e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.460s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.294465] env[61857]: DEBUG nova.network.neutron [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.499641] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.079s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.500460] env[61857]: ERROR nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d9602eb9-b5bf-4cf1-a971-2ca9191b94d9, please check neutron logs for more information. [ 566.500460] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Traceback (most recent call last): [ 566.500460] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.500460] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] self.driver.spawn(context, instance, image_meta, [ 566.500460] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 566.500460] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.500460] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.500460] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] vm_ref = self.build_virtual_machine(instance, [ 566.500460] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.500460] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.500460] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] for vif in network_info: [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] return self._sync_wrapper(fn, *args, **kwargs) [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] self.wait() [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] self[:] = self._gt.wait() [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] return self._exit_event.wait() [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] result = hub.switch() [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.500863] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] return self.greenlet.switch() [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] result = function(*args, **kwargs) [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] return func(*args, **kwargs) [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] raise e [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] nwinfo = self.network_api.allocate_for_instance( [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] created_port_ids = self._update_ports_for_instance( [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] with excutils.save_and_reraise_exception(): [ 566.501295] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] self.force_reraise() [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] raise self.value [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] updated_port = self._update_port( [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] _ensure_no_port_binding_failure(port) [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] raise exception.PortBindingFailed(port_id=port['id']) [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] nova.exception.PortBindingFailed: Binding failed for port d9602eb9-b5bf-4cf1-a971-2ca9191b94d9, please check neutron logs for more information. [ 566.501669] env[61857]: ERROR nova.compute.manager [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] [ 566.501951] env[61857]: DEBUG nova.compute.utils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Binding failed for port d9602eb9-b5bf-4cf1-a971-2ca9191b94d9, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 566.502425] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.757s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.502607] env[61857]: DEBUG nova.objects.instance [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61857) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 566.509253] env[61857]: DEBUG nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Build of instance 67ff728e-37b1-4633-bfd2-75fe4b1c7d13 was re-scheduled: Binding failed for port d9602eb9-b5bf-4cf1-a971-2ca9191b94d9, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 566.511337] env[61857]: DEBUG nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 566.511337] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "refresh_cache-67ff728e-37b1-4633-bfd2-75fe4b1c7d13" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.511337] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "refresh_cache-67ff728e-37b1-4633-bfd2-75fe4b1c7d13" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.511337] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 566.548348] env[61857]: DEBUG nova.network.neutron [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.625962] env[61857]: DEBUG nova.network.neutron [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.673664] env[61857]: DEBUG nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 566.797375] env[61857]: INFO nova.compute.manager [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] [instance: 4c7dab9e-4527-46b6-a1ec-6e6935768b75] Took 1.05 seconds to deallocate network for instance. [ 566.945605] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "249ffa46-c5c7-4669-84a5-c02e0799dc59" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.945965] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "249ffa46-c5c7-4669-84a5-c02e0799dc59" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.037778] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.129096] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Releasing lock "refresh_cache-bb418d62-d47c-42c5-908e-26f1eb034e46" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.129524] env[61857]: DEBUG nova.compute.manager [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 567.129752] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 567.132195] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4a43b0-6d25-4d7a-b5c0-8ce04d2cd382 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.139123] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 567.139384] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9dd5fd5b-b7cf-40e0-a830-141bda0aedd5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.146463] env[61857]: DEBUG oslo_vmware.api [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 567.146463] env[61857]: value = "task-950701" [ 567.146463] env[61857]: _type = "Task" [ 567.146463] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.156752] env[61857]: DEBUG oslo_vmware.api [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950701, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.204918] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.206452] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.518906] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4a862f6e-197e-41ae-a41d-2cea12b70473 tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.521120] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.391s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.526256] env[61857]: INFO nova.compute.claims [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 567.666877] env[61857]: DEBUG oslo_vmware.api [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950701, 'name': PowerOffVM_Task, 'duration_secs': 0.240992} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.666877] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 567.666877] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 567.669513] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59b39922-21e3-4430-8e16-97af9bbce6d8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.692638] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 567.692857] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 567.693037] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleting the datastore file [datastore2] bb418d62-d47c-42c5-908e-26f1eb034e46 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 567.693437] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75a4edc5-cb90-43d0-8848-94e0a49efaaa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.704421] env[61857]: DEBUG oslo_vmware.api [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for the task: (returnval){ [ 567.704421] env[61857]: value = "task-950703" [ 567.704421] env[61857]: _type = "Task" [ 567.704421] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.709135] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "refresh_cache-67ff728e-37b1-4633-bfd2-75fe4b1c7d13" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.709358] env[61857]: DEBUG nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 567.709534] env[61857]: DEBUG nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 567.709708] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 567.721167] env[61857]: DEBUG oslo_vmware.api [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950703, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.755683] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.846673] env[61857]: INFO nova.scheduler.client.report [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Deleted allocations for instance 4c7dab9e-4527-46b6-a1ec-6e6935768b75 [ 568.216590] env[61857]: DEBUG oslo_vmware.api [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Task: {'id': task-950703, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164844} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.216873] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 568.216963] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 568.220246] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 568.220246] env[61857]: INFO nova.compute.manager [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Took 1.09 seconds to destroy the instance on the hypervisor. [ 568.220246] env[61857]: DEBUG oslo.service.loopingcall [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 568.220246] env[61857]: DEBUG nova.compute.manager [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 568.220246] env[61857]: DEBUG nova.network.neutron [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 568.250343] env[61857]: DEBUG nova.network.neutron [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.257878] env[61857]: DEBUG nova.network.neutron [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.355711] env[61857]: DEBUG oslo_concurrency.lockutils [None req-020d2bb8-f018-4e85-b864-c918238560cc tempest-ServerExternalEventsTest-999407631 tempest-ServerExternalEventsTest-999407631-project-member] Lock "4c7dab9e-4527-46b6-a1ec-6e6935768b75" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.285s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.758977] env[61857]: DEBUG nova.network.neutron [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.763266] env[61857]: INFO nova.compute.manager [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 67ff728e-37b1-4633-bfd2-75fe4b1c7d13] Took 1.05 seconds to deallocate network for instance. [ 568.866554] env[61857]: DEBUG nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 569.024338] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f409148c-e482-4fd3-b075-7f6697745eff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.036296] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207f6c75-b8c5-445c-9c12-0536b0ea5166 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.072873] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d64390-22f3-486f-b70e-137d51cde4ce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.083250] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09775475-615c-4f6f-8803-25e623a76496 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.100414] env[61857]: DEBUG nova.compute.provider_tree [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.261237] env[61857]: INFO nova.compute.manager [-] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Took 1.04 seconds to deallocate network for instance. [ 569.399796] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.602062] env[61857]: DEBUG nova.scheduler.client.report [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.769364] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.806303] env[61857]: INFO nova.scheduler.client.report [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleted allocations for instance 67ff728e-37b1-4633-bfd2-75fe4b1c7d13 [ 570.109450] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.111591] env[61857]: DEBUG nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.115382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.741s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.116969] env[61857]: INFO nova.compute.claims [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 570.316855] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3d7ff237-44ed-4a06-b7e8-71387440a7b3 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "67ff728e-37b1-4633-bfd2-75fe4b1c7d13" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.973s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.622561] env[61857]: DEBUG nova.compute.utils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 570.626051] env[61857]: DEBUG nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 570.626239] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 570.725517] env[61857]: DEBUG nova.policy [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '78c9e02d2f4b428bab1aa09f0e27c827', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7b20cbce6494b21b660e9a37b582767', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 570.821199] env[61857]: DEBUG nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 571.126709] env[61857]: DEBUG nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.352317] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.484848] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Successfully created port: 40296f2c-b810-427f-bceb-3b149c9f88e1 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 571.539720] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5a84d8-69d5-45de-878b-6a4c4549e276 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.549014] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90f1067-f88e-446e-ba1c-255deab7d04f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.599981] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6bc401-835c-4a0f-901e-6c9bb6ef34c8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.612491] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b995a15-63fd-4592-8238-7b473791c186 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.644183] env[61857]: DEBUG nova.compute.provider_tree [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.147157] env[61857]: DEBUG nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.151791] env[61857]: DEBUG nova.scheduler.client.report [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.165884] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Acquiring lock "95fbef26-78a7-4449-ba06-ffd517fb1032" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.166081] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Lock "95fbef26-78a7-4449-ba06-ffd517fb1032" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.178804] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.179009] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.179191] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.179367] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.179512] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.179659] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.179860] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.180029] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.180689] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.180689] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.180689] env[61857]: DEBUG nova.virt.hardware [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.181970] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678c850a-c5fc-4fa7-9851-55e75bcc00f3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.189769] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c780a70-bab5-4cac-8497-ffe755649255 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.656976] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.657548] env[61857]: DEBUG nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 572.665020] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.832s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.053142] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Acquiring lock "4ede4082-93b0-4cac-9e28-5448f04cba9b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.053422] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Lock "4ede4082-93b0-4cac-9e28-5448f04cba9b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.169247] env[61857]: DEBUG nova.compute.utils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 573.182698] env[61857]: DEBUG nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 573.182698] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 573.316681] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "51c54fe3-c0e2-4151-9626-cfb10cd7996e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.316946] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "51c54fe3-c0e2-4151-9626-cfb10cd7996e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.356063] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "8ee206cf-c906-4ac0-8ba7-d864f5f538db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.356063] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "8ee206cf-c906-4ac0-8ba7-d864f5f538db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.402352] env[61857]: DEBUG nova.policy [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f5af5f69b9cb428d9453bc023e16aeaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '18397efabee24a42b6eb6021ab7a04b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 573.480584] env[61857]: ERROR nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 40296f2c-b810-427f-bceb-3b149c9f88e1, please check neutron logs for more information. [ 573.480584] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 573.480584] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.480584] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 573.480584] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.480584] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 573.480584] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.480584] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 573.480584] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.480584] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 573.480584] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.480584] env[61857]: ERROR nova.compute.manager raise self.value [ 573.480584] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.480584] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 573.480584] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.480584] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 573.481109] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.481109] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 573.481109] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 40296f2c-b810-427f-bceb-3b149c9f88e1, please check neutron logs for more information. [ 573.481109] env[61857]: ERROR nova.compute.manager [ 573.481109] env[61857]: Traceback (most recent call last): [ 573.481109] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 573.481109] env[61857]: listener.cb(fileno) [ 573.481109] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.481109] env[61857]: result = function(*args, **kwargs) [ 573.481109] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.481109] env[61857]: return func(*args, **kwargs) [ 573.481109] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.481109] env[61857]: raise e [ 573.481109] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.481109] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 573.481109] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.481109] env[61857]: created_port_ids = self._update_ports_for_instance( [ 573.481109] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.481109] env[61857]: with excutils.save_and_reraise_exception(): [ 573.481109] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.481109] env[61857]: self.force_reraise() [ 573.481109] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.481109] env[61857]: raise self.value [ 573.481109] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.481109] env[61857]: updated_port = self._update_port( [ 573.481109] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.481109] env[61857]: _ensure_no_port_binding_failure(port) [ 573.481109] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.481109] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 573.481883] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 40296f2c-b810-427f-bceb-3b149c9f88e1, please check neutron logs for more information. [ 573.481883] env[61857]: Removing descriptor: 17 [ 573.481883] env[61857]: ERROR nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 40296f2c-b810-427f-bceb-3b149c9f88e1, please check neutron logs for more information. [ 573.481883] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Traceback (most recent call last): [ 573.481883] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 573.481883] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] yield resources [ 573.481883] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.481883] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] self.driver.spawn(context, instance, image_meta, [ 573.481883] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.481883] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.481883] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.481883] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] vm_ref = self.build_virtual_machine(instance, [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] for vif in network_info: [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] return self._sync_wrapper(fn, *args, **kwargs) [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] self.wait() [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] self[:] = self._gt.wait() [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] return self._exit_event.wait() [ 573.482229] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] result = hub.switch() [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] return self.greenlet.switch() [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] result = function(*args, **kwargs) [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] return func(*args, **kwargs) [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] raise e [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] nwinfo = self.network_api.allocate_for_instance( [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.482604] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] created_port_ids = self._update_ports_for_instance( [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] with excutils.save_and_reraise_exception(): [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] self.force_reraise() [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] raise self.value [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] updated_port = self._update_port( [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] _ensure_no_port_binding_failure(port) [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.482966] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] raise exception.PortBindingFailed(port_id=port['id']) [ 573.483328] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] nova.exception.PortBindingFailed: Binding failed for port 40296f2c-b810-427f-bceb-3b149c9f88e1, please check neutron logs for more information. [ 573.483328] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] [ 573.483328] env[61857]: INFO nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Terminating instance [ 573.483328] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Acquiring lock "refresh_cache-f38e8137-da78-40f7-88a1-0f870a83ffd9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.483328] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Acquired lock "refresh_cache-f38e8137-da78-40f7-88a1-0f870a83ffd9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.483328] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.609479] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22095d6d-256b-421e-96f3-791bdc8f557b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.619534] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33ae4697-edcb-494f-9414-63da9ba1957b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.647707] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611d7ba5-0ea5-461c-9e40-5f91239d7fe3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.656748] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72d8a2b-98a0-49f3-a32b-6e171556d53e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.670519] env[61857]: DEBUG nova.compute.provider_tree [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.681042] env[61857]: DEBUG nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 574.003455] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.028132] env[61857]: DEBUG nova.compute.manager [req-589c0227-2f8e-4d93-b7cb-69334a79952a req-ab612d79-a343-4144-b8a9-4d88bb03c688 service nova] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Received event network-changed-40296f2c-b810-427f-bceb-3b149c9f88e1 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 574.028132] env[61857]: DEBUG nova.compute.manager [req-589c0227-2f8e-4d93-b7cb-69334a79952a req-ab612d79-a343-4144-b8a9-4d88bb03c688 service nova] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Refreshing instance network info cache due to event network-changed-40296f2c-b810-427f-bceb-3b149c9f88e1. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 574.028132] env[61857]: DEBUG oslo_concurrency.lockutils [req-589c0227-2f8e-4d93-b7cb-69334a79952a req-ab612d79-a343-4144-b8a9-4d88bb03c688 service nova] Acquiring lock "refresh_cache-f38e8137-da78-40f7-88a1-0f870a83ffd9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.137567] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.173105] env[61857]: DEBUG nova.scheduler.client.report [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.192848] env[61857]: DEBUG nova.compute.manager [None req-b5cf4fc4-fe66-44d2-a538-09da84cf1b03 tempest-ServerDiagnosticsV248Test-2033239442 tempest-ServerDiagnosticsV248Test-2033239442-project-admin] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 574.195072] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b17992-8af2-4ab2-964e-7d8fad74f1da {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.209270] env[61857]: INFO nova.compute.manager [None req-b5cf4fc4-fe66-44d2-a538-09da84cf1b03 tempest-ServerDiagnosticsV248Test-2033239442 tempest-ServerDiagnosticsV248Test-2033239442-project-admin] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Retrieving diagnostics [ 574.210222] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84109055-a18f-4536-9278-d2b34e2016c8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.640347] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Releasing lock "refresh_cache-f38e8137-da78-40f7-88a1-0f870a83ffd9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.640782] env[61857]: DEBUG nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 574.640972] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 574.641335] env[61857]: DEBUG oslo_concurrency.lockutils [req-589c0227-2f8e-4d93-b7cb-69334a79952a req-ab612d79-a343-4144-b8a9-4d88bb03c688 service nova] Acquired lock "refresh_cache-f38e8137-da78-40f7-88a1-0f870a83ffd9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.641509] env[61857]: DEBUG nova.network.neutron [req-589c0227-2f8e-4d93-b7cb-69334a79952a req-ab612d79-a343-4144-b8a9-4d88bb03c688 service nova] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Refreshing network info cache for port 40296f2c-b810-427f-bceb-3b149c9f88e1 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 574.642575] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2d666e5-c57d-428e-bc4a-3fff56256749 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.659921] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c37513d-ace9-4a3b-b347-caa6f2b4a8da {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.677514] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.017s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.678213] env[61857]: ERROR nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d54775be-555a-4588-8282-22e67b6b3d2a, please check neutron logs for more information. [ 574.678213] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Traceback (most recent call last): [ 574.678213] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 574.678213] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] self.driver.spawn(context, instance, image_meta, [ 574.678213] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 574.678213] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.678213] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.678213] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] vm_ref = self.build_virtual_machine(instance, [ 574.678213] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.678213] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.678213] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] for vif in network_info: [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] return self._sync_wrapper(fn, *args, **kwargs) [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] self.wait() [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] self[:] = self._gt.wait() [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] return self._exit_event.wait() [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] result = hub.switch() [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.678591] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] return self.greenlet.switch() [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] result = function(*args, **kwargs) [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] return func(*args, **kwargs) [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] raise e [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] nwinfo = self.network_api.allocate_for_instance( [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] created_port_ids = self._update_ports_for_instance( [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] with excutils.save_and_reraise_exception(): [ 574.678963] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] self.force_reraise() [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] raise self.value [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] updated_port = self._update_port( [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] _ensure_no_port_binding_failure(port) [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] raise exception.PortBindingFailed(port_id=port['id']) [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] nova.exception.PortBindingFailed: Binding failed for port d54775be-555a-4588-8282-22e67b6b3d2a, please check neutron logs for more information. [ 574.679399] env[61857]: ERROR nova.compute.manager [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] [ 574.680031] env[61857]: DEBUG nova.compute.utils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Binding failed for port d54775be-555a-4588-8282-22e67b6b3d2a, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 574.680734] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.910s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.682616] env[61857]: INFO nova.compute.claims [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 574.688176] env[61857]: DEBUG nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Build of instance ad56d32b-c78d-45db-872a-1c4ea91ab909 was re-scheduled: Binding failed for port d54775be-555a-4588-8282-22e67b6b3d2a, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 574.688176] env[61857]: DEBUG nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 574.688176] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Acquiring lock "refresh_cache-ad56d32b-c78d-45db-872a-1c4ea91ab909" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.688176] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Acquired lock "refresh_cache-ad56d32b-c78d-45db-872a-1c4ea91ab909" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.688347] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 574.693099] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f38e8137-da78-40f7-88a1-0f870a83ffd9 could not be found. [ 574.693327] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 574.693528] env[61857]: INFO nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 574.694039] env[61857]: DEBUG oslo.service.loopingcall [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.695539] env[61857]: DEBUG nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 574.700813] env[61857]: DEBUG nova.compute.manager [-] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.700813] env[61857]: DEBUG nova.network.neutron [-] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.702885] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Successfully created port: 53eb146e-3d62-4715-b230-6f04afc7b67c {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.725339] env[61857]: DEBUG nova.network.neutron [-] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.733361] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 574.733742] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 574.734091] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.734346] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 574.734565] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.735137] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 574.735728] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 574.735988] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 574.736246] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 574.736470] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 574.736691] env[61857]: DEBUG nova.virt.hardware [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 574.738418] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f2c11d-bccc-45bc-ba49-d1bb46ef0a05 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.748548] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c36865-2520-422b-81a0-af402f589173 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.202398] env[61857]: DEBUG nova.network.neutron [req-589c0227-2f8e-4d93-b7cb-69334a79952a req-ab612d79-a343-4144-b8a9-4d88bb03c688 service nova] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.229605] env[61857]: DEBUG nova.network.neutron [-] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.259132] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.464718] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "ca3d3ad9-39a0-467b-812f-21b7303b8bc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.464961] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "ca3d3ad9-39a0-467b-812f-21b7303b8bc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.497717] env[61857]: DEBUG nova.network.neutron [req-589c0227-2f8e-4d93-b7cb-69334a79952a req-ab612d79-a343-4144-b8a9-4d88bb03c688 service nova] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.517919] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.734776] env[61857]: INFO nova.compute.manager [-] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Took 1.04 seconds to deallocate network for instance. [ 575.737551] env[61857]: DEBUG nova.compute.claims [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.738045] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.000350] env[61857]: DEBUG oslo_concurrency.lockutils [req-589c0227-2f8e-4d93-b7cb-69334a79952a req-ab612d79-a343-4144-b8a9-4d88bb03c688 service nova] Releasing lock "refresh_cache-f38e8137-da78-40f7-88a1-0f870a83ffd9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.020387] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Releasing lock "refresh_cache-ad56d32b-c78d-45db-872a-1c4ea91ab909" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.020502] env[61857]: DEBUG nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 576.020644] env[61857]: DEBUG nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.020786] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 576.081264] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.155793] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquiring lock "2e7c154b-dab2-4384-bdb8-d098967654c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.155793] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lock "2e7c154b-dab2-4384-bdb8-d098967654c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.155793] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquiring lock "2e7c154b-dab2-4384-bdb8-d098967654c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.155793] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lock "2e7c154b-dab2-4384-bdb8-d098967654c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.155997] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lock "2e7c154b-dab2-4384-bdb8-d098967654c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.159129] env[61857]: INFO nova.compute.manager [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Terminating instance [ 576.162643] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquiring lock "refresh_cache-2e7c154b-dab2-4384-bdb8-d098967654c5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.164205] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquired lock "refresh_cache-2e7c154b-dab2-4384-bdb8-d098967654c5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.164205] env[61857]: DEBUG nova.network.neutron [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 576.234568] env[61857]: DEBUG nova.compute.manager [req-39769dfe-a6b4-47de-938d-af77fe7dd491 req-c684ed35-9166-408e-8280-681aa717a292 service nova] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Received event network-vif-deleted-40296f2c-b810-427f-bceb-3b149c9f88e1 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 576.255552] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a731c6-8f86-4de2-b8d2-2d4a656d3ff0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.265687] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33c7140-48d7-4ab1-a4ed-c965de3ef1b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.307870] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead30011-9931-4381-ac55-1926f29be2e1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.316781] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e290189-640c-4b71-84a4-1ef13e2cbc91 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.330035] env[61857]: DEBUG nova.compute.provider_tree [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.585341] env[61857]: DEBUG nova.network.neutron [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.720248] env[61857]: DEBUG nova.network.neutron [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.833788] env[61857]: DEBUG nova.scheduler.client.report [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.908296] env[61857]: DEBUG nova.network.neutron [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.087210] env[61857]: INFO nova.compute.manager [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] [instance: ad56d32b-c78d-45db-872a-1c4ea91ab909] Took 1.07 seconds to deallocate network for instance. [ 577.340229] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.659s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.341026] env[61857]: DEBUG nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.347974] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.630s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.350104] env[61857]: INFO nova.compute.claims [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.413505] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Releasing lock "refresh_cache-2e7c154b-dab2-4384-bdb8-d098967654c5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.413505] env[61857]: DEBUG nova.compute.manager [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 577.413505] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 577.413505] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb383dbf-dcb9-486c-98c0-6d8cd42a7f4a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.424715] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 577.426203] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e398a58-066a-4bf8-9057-27ca90c98949 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.434937] env[61857]: DEBUG oslo_vmware.api [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for the task: (returnval){ [ 577.434937] env[61857]: value = "task-950709" [ 577.434937] env[61857]: _type = "Task" [ 577.434937] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.455338] env[61857]: DEBUG oslo_vmware.api [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950709, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.587197] env[61857]: DEBUG nova.compute.manager [req-589961dd-5f83-4c5a-bf23-c926338791f3 req-fd4e2310-aab0-4354-8da3-82cbfc080de8 service nova] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Received event network-changed-53eb146e-3d62-4715-b230-6f04afc7b67c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 577.587197] env[61857]: DEBUG nova.compute.manager [req-589961dd-5f83-4c5a-bf23-c926338791f3 req-fd4e2310-aab0-4354-8da3-82cbfc080de8 service nova] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Refreshing instance network info cache due to event network-changed-53eb146e-3d62-4715-b230-6f04afc7b67c. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 577.587197] env[61857]: DEBUG oslo_concurrency.lockutils [req-589961dd-5f83-4c5a-bf23-c926338791f3 req-fd4e2310-aab0-4354-8da3-82cbfc080de8 service nova] Acquiring lock "refresh_cache-0b6029b8-3267-4662-a1af-8e0b6af3288f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.587197] env[61857]: DEBUG oslo_concurrency.lockutils [req-589961dd-5f83-4c5a-bf23-c926338791f3 req-fd4e2310-aab0-4354-8da3-82cbfc080de8 service nova] Acquired lock "refresh_cache-0b6029b8-3267-4662-a1af-8e0b6af3288f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.587197] env[61857]: DEBUG nova.network.neutron [req-589961dd-5f83-4c5a-bf23-c926338791f3 req-fd4e2310-aab0-4354-8da3-82cbfc080de8 service nova] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Refreshing network info cache for port 53eb146e-3d62-4715-b230-6f04afc7b67c {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 577.855296] env[61857]: DEBUG nova.compute.utils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 577.859562] env[61857]: DEBUG nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 577.859812] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 577.954679] env[61857]: DEBUG oslo_vmware.api [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950709, 'name': PowerOffVM_Task, 'duration_secs': 0.116243} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.955578] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 577.956362] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 577.956985] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a53e35cf-34b3-4407-955c-c1ba598aeca9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.983061] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 577.983329] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 577.983524] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Deleting the datastore file [datastore2] 2e7c154b-dab2-4384-bdb8-d098967654c5 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 577.983778] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f1438857-6fbe-4618-819a-2076b4595141 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.990193] env[61857]: DEBUG oslo_vmware.api [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for the task: (returnval){ [ 577.990193] env[61857]: value = "task-950711" [ 577.990193] env[61857]: _type = "Task" [ 577.990193] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.998949] env[61857]: DEBUG oslo_vmware.api [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950711, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.000682] env[61857]: DEBUG nova.policy [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45523fc9fdd04ad9859e73e4e1508d8a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c0f687f57323442c993fc2e52c9ea790', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 578.137223] env[61857]: INFO nova.scheduler.client.report [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Deleted allocations for instance ad56d32b-c78d-45db-872a-1c4ea91ab909 [ 578.151623] env[61857]: ERROR nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 53eb146e-3d62-4715-b230-6f04afc7b67c, please check neutron logs for more information. [ 578.151623] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 578.151623] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.151623] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 578.151623] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 578.151623] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 578.151623] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 578.151623] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 578.151623] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.151623] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 578.151623] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.151623] env[61857]: ERROR nova.compute.manager raise self.value [ 578.151623] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 578.151623] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 578.151623] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.151623] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 578.152222] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.152222] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 578.152222] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 53eb146e-3d62-4715-b230-6f04afc7b67c, please check neutron logs for more information. [ 578.152222] env[61857]: ERROR nova.compute.manager [ 578.152222] env[61857]: Traceback (most recent call last): [ 578.152222] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 578.152222] env[61857]: listener.cb(fileno) [ 578.152222] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.152222] env[61857]: result = function(*args, **kwargs) [ 578.152222] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 578.152222] env[61857]: return func(*args, **kwargs) [ 578.152222] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.152222] env[61857]: raise e [ 578.152222] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.152222] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 578.152222] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 578.152222] env[61857]: created_port_ids = self._update_ports_for_instance( [ 578.152222] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 578.152222] env[61857]: with excutils.save_and_reraise_exception(): [ 578.152222] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.152222] env[61857]: self.force_reraise() [ 578.152222] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.152222] env[61857]: raise self.value [ 578.152222] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 578.152222] env[61857]: updated_port = self._update_port( [ 578.152222] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.152222] env[61857]: _ensure_no_port_binding_failure(port) [ 578.152222] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.152222] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 578.153105] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 53eb146e-3d62-4715-b230-6f04afc7b67c, please check neutron logs for more information. [ 578.153105] env[61857]: Removing descriptor: 16 [ 578.153105] env[61857]: ERROR nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 53eb146e-3d62-4715-b230-6f04afc7b67c, please check neutron logs for more information. [ 578.153105] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Traceback (most recent call last): [ 578.153105] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 578.153105] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] yield resources [ 578.153105] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 578.153105] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] self.driver.spawn(context, instance, image_meta, [ 578.153105] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 578.153105] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.153105] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.153105] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] vm_ref = self.build_virtual_machine(instance, [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] for vif in network_info: [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] return self._sync_wrapper(fn, *args, **kwargs) [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] self.wait() [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] self[:] = self._gt.wait() [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] return self._exit_event.wait() [ 578.153483] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] result = hub.switch() [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] return self.greenlet.switch() [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] result = function(*args, **kwargs) [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] return func(*args, **kwargs) [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] raise e [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] nwinfo = self.network_api.allocate_for_instance( [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 578.154203] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] created_port_ids = self._update_ports_for_instance( [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] with excutils.save_and_reraise_exception(): [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] self.force_reraise() [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] raise self.value [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] updated_port = self._update_port( [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] _ensure_no_port_binding_failure(port) [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.154646] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] raise exception.PortBindingFailed(port_id=port['id']) [ 578.156147] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] nova.exception.PortBindingFailed: Binding failed for port 53eb146e-3d62-4715-b230-6f04afc7b67c, please check neutron logs for more information. [ 578.156147] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] [ 578.156147] env[61857]: INFO nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Terminating instance [ 578.157843] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquiring lock "refresh_cache-0b6029b8-3267-4662-a1af-8e0b6af3288f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.194240] env[61857]: DEBUG nova.network.neutron [req-589961dd-5f83-4c5a-bf23-c926338791f3 req-fd4e2310-aab0-4354-8da3-82cbfc080de8 service nova] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.364028] env[61857]: DEBUG nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.503933] env[61857]: DEBUG oslo_vmware.api [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Task: {'id': task-950711, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09468} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.505139] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 578.505139] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 578.505139] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 578.505139] env[61857]: INFO nova.compute.manager [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Took 1.09 seconds to destroy the instance on the hypervisor. [ 578.505139] env[61857]: DEBUG oslo.service.loopingcall [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 578.505345] env[61857]: DEBUG nova.compute.manager [-] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.505345] env[61857]: DEBUG nova.network.neutron [-] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 578.528890] env[61857]: DEBUG nova.network.neutron [-] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.647030] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3f6bf393-8263-4060-bf76-f5fbc596799f tempest-ServerActionsTestOtherA-1321980693 tempest-ServerActionsTestOtherA-1321980693-project-member] Lock "ad56d32b-c78d-45db-872a-1c4ea91ab909" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.548s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.719245] env[61857]: DEBUG nova.network.neutron [req-589961dd-5f83-4c5a-bf23-c926338791f3 req-fd4e2310-aab0-4354-8da3-82cbfc080de8 service nova] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.857297] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42cc96c8-4d11-40c7-bcae-7d347905f264 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.868391] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f504b4-f2ca-4db4-858d-37e45b759c89 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.911150] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-734c317e-5f65-4df2-8b1f-90950c19ea68 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.920482] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f8fee8-6231-4794-ac68-3c178c9f9044 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.942131] env[61857]: DEBUG nova.compute.provider_tree [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.035084] env[61857]: DEBUG nova.network.neutron [-] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.150222] env[61857]: DEBUG nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.222286] env[61857]: DEBUG oslo_concurrency.lockutils [req-589961dd-5f83-4c5a-bf23-c926338791f3 req-fd4e2310-aab0-4354-8da3-82cbfc080de8 service nova] Releasing lock "refresh_cache-0b6029b8-3267-4662-a1af-8e0b6af3288f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.225569] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquired lock "refresh_cache-0b6029b8-3267-4662-a1af-8e0b6af3288f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.225569] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 579.381827] env[61857]: DEBUG nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.412981] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:38:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1828454556',id=24,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1492388455',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.413271] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.413432] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.413616] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.413763] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.413912] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.414604] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.414825] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.415027] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.415428] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.415658] env[61857]: DEBUG nova.virt.hardware [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.416587] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1db793-102e-4f1d-8937-878677465ca5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.426860] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ad6728-a162-4f6b-93c0-20eab3569b52 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.446197] env[61857]: DEBUG nova.scheduler.client.report [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.538099] env[61857]: INFO nova.compute.manager [-] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Took 1.03 seconds to deallocate network for instance. [ 579.699785] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.766938] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.770893] env[61857]: DEBUG nova.compute.manager [req-94d90786-e32d-4541-ac1d-b9f3793928d9 req-0264d738-242d-4cbc-8049-d17903c91778 service nova] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Received event network-vif-deleted-53eb146e-3d62-4715-b230-6f04afc7b67c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 579.771785] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Successfully created port: 2a9e5e0b-b590-4ea0-882a-c281c286e7c9 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.959019] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.610s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.959019] env[61857]: DEBUG nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 579.963621] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.240s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.963818] env[61857]: DEBUG nova.objects.instance [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61857) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 580.055025] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.064500] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.476021] env[61857]: DEBUG nova.compute.utils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.480597] env[61857]: DEBUG nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 580.481176] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 580.571018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Releasing lock "refresh_cache-0b6029b8-3267-4662-a1af-8e0b6af3288f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.571018] env[61857]: DEBUG nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.571018] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 580.571018] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5be9c64e-4b09-48c1-ad6f-47941dcec910 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.579021] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c295ee-8681-4e7c-9e28-3f0e7c5086bd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.608157] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0b6029b8-3267-4662-a1af-8e0b6af3288f could not be found. [ 580.608157] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 580.608157] env[61857]: INFO nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 580.608157] env[61857]: DEBUG oslo.service.loopingcall [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.608157] env[61857]: DEBUG nova.compute.manager [-] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.608157] env[61857]: DEBUG nova.network.neutron [-] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 580.644303] env[61857]: DEBUG nova.network.neutron [-] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.660212] env[61857]: DEBUG nova.policy [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67e27d2b59c74401a6141014441d3cd7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e3d3d12782a4a0fbdff54566d58cfa5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.984216] env[61857]: DEBUG nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 580.985950] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4b6569ee-a8a5-47cf-8e26-8a777e10331c tempest-ServersAdmin275Test-1525437356 tempest-ServersAdmin275Test-1525437356-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.987269] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.945s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.990343] env[61857]: INFO nova.compute.claims [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 581.148486] env[61857]: DEBUG nova.network.neutron [-] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.507977] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "9edf9e8c-5421-4018-afee-59e10cd335ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.509104] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "9edf9e8c-5421-4018-afee-59e10cd335ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.652185] env[61857]: INFO nova.compute.manager [-] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Took 1.04 seconds to deallocate network for instance. [ 581.653413] env[61857]: DEBUG nova.compute.claims [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.653413] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.995424] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Successfully created port: 715f7af3-9869-4bfa-a356-6e4d4920728d {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 581.998023] env[61857]: DEBUG nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 582.049348] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.049348] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.049348] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.049694] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.049862] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.050022] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.050278] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.050462] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.050671] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.050842] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.051029] env[61857]: DEBUG nova.virt.hardware [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.052009] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c966baa2-794c-4b94-87d1-3582a32e3a07 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.064367] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d42810d-4d29-47eb-b74e-36fb8f09fb22 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.366890] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Acquiring lock "e52972ca-877f-45b8-b79f-c834246a8857" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.367673] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Lock "e52972ca-877f-45b8-b79f-c834246a8857" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.513225] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6a8b42-ff84-4787-a835-bb35dc18844f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.521800] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb446b4-2ec7-4942-8692-f6a2c71efcbb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.562533] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d5f1b5-069a-43fd-af78-585d2d79ca1e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.571341] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd38a1e-aca4-4829-a284-1d29944126a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.585430] env[61857]: DEBUG nova.compute.provider_tree [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.092928] env[61857]: DEBUG nova.scheduler.client.report [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 583.596634] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.597252] env[61857]: DEBUG nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 583.603762] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.614s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.108594] env[61857]: DEBUG nova.compute.utils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 584.110696] env[61857]: DEBUG nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 584.110696] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 584.206167] env[61857]: DEBUG nova.compute.manager [req-9780af18-208c-43e0-80b0-55e2058f8e78 req-183503ff-c192-4795-b65b-25a9131209c3 service nova] [instance: e7e82254-4088-4232-b861-9aee864c3737] Received event network-changed-2a9e5e0b-b590-4ea0-882a-c281c286e7c9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 584.206388] env[61857]: DEBUG nova.compute.manager [req-9780af18-208c-43e0-80b0-55e2058f8e78 req-183503ff-c192-4795-b65b-25a9131209c3 service nova] [instance: e7e82254-4088-4232-b861-9aee864c3737] Refreshing instance network info cache due to event network-changed-2a9e5e0b-b590-4ea0-882a-c281c286e7c9. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 584.206629] env[61857]: DEBUG oslo_concurrency.lockutils [req-9780af18-208c-43e0-80b0-55e2058f8e78 req-183503ff-c192-4795-b65b-25a9131209c3 service nova] Acquiring lock "refresh_cache-e7e82254-4088-4232-b861-9aee864c3737" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.206771] env[61857]: DEBUG oslo_concurrency.lockutils [req-9780af18-208c-43e0-80b0-55e2058f8e78 req-183503ff-c192-4795-b65b-25a9131209c3 service nova] Acquired lock "refresh_cache-e7e82254-4088-4232-b861-9aee864c3737" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.206928] env[61857]: DEBUG nova.network.neutron [req-9780af18-208c-43e0-80b0-55e2058f8e78 req-183503ff-c192-4795-b65b-25a9131209c3 service nova] [instance: e7e82254-4088-4232-b861-9aee864c3737] Refreshing network info cache for port 2a9e5e0b-b590-4ea0-882a-c281c286e7c9 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 584.268985] env[61857]: DEBUG nova.policy [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd169a99b1b904c589b4a062de88b5c6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a845df4e867940639f4ff7b6a092e301', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 584.580178] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c79b80e-29d4-437f-86af-d16b326aa332 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.592019] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4e51b7-e7e5-42ed-b3ed-7e54328c8b6a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.624558] env[61857]: DEBUG nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 584.633665] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e85061d-b3ea-466a-8b05-b7bfc48f6c73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.644264] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725bdd9a-348d-4962-9de3-f93d99f454d8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.659700] env[61857]: DEBUG nova.compute.provider_tree [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.878691] env[61857]: DEBUG nova.network.neutron [req-9780af18-208c-43e0-80b0-55e2058f8e78 req-183503ff-c192-4795-b65b-25a9131209c3 service nova] [instance: e7e82254-4088-4232-b861-9aee864c3737] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.164936] env[61857]: DEBUG nova.scheduler.client.report [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.487288] env[61857]: DEBUG nova.network.neutron [req-9780af18-208c-43e0-80b0-55e2058f8e78 req-183503ff-c192-4795-b65b-25a9131209c3 service nova] [instance: e7e82254-4088-4232-b861-9aee864c3737] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.495646] env[61857]: ERROR nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2a9e5e0b-b590-4ea0-882a-c281c286e7c9, please check neutron logs for more information. [ 585.495646] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.495646] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.495646] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.495646] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.495646] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.495646] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.495646] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.495646] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.495646] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 585.495646] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.495646] env[61857]: ERROR nova.compute.manager raise self.value [ 585.495646] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.495646] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.495646] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.495646] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.496156] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.496156] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.496156] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2a9e5e0b-b590-4ea0-882a-c281c286e7c9, please check neutron logs for more information. [ 585.496156] env[61857]: ERROR nova.compute.manager [ 585.496156] env[61857]: Traceback (most recent call last): [ 585.496156] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.496156] env[61857]: listener.cb(fileno) [ 585.496156] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.496156] env[61857]: result = function(*args, **kwargs) [ 585.496156] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.496156] env[61857]: return func(*args, **kwargs) [ 585.496156] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.496156] env[61857]: raise e [ 585.496156] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.496156] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 585.496156] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.496156] env[61857]: created_port_ids = self._update_ports_for_instance( [ 585.496156] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.496156] env[61857]: with excutils.save_and_reraise_exception(): [ 585.496156] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.496156] env[61857]: self.force_reraise() [ 585.496156] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.496156] env[61857]: raise self.value [ 585.496156] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.496156] env[61857]: updated_port = self._update_port( [ 585.496156] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.496156] env[61857]: _ensure_no_port_binding_failure(port) [ 585.496156] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.496156] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.496913] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 2a9e5e0b-b590-4ea0-882a-c281c286e7c9, please check neutron logs for more information. [ 585.496913] env[61857]: Removing descriptor: 20 [ 585.496913] env[61857]: ERROR nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2a9e5e0b-b590-4ea0-882a-c281c286e7c9, please check neutron logs for more information. [ 585.496913] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] Traceback (most recent call last): [ 585.496913] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 585.496913] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] yield resources [ 585.496913] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.496913] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] self.driver.spawn(context, instance, image_meta, [ 585.496913] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 585.496913] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.496913] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.496913] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] vm_ref = self.build_virtual_machine(instance, [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] for vif in network_info: [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] return self._sync_wrapper(fn, *args, **kwargs) [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] self.wait() [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] self[:] = self._gt.wait() [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] return self._exit_event.wait() [ 585.497259] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] result = hub.switch() [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] return self.greenlet.switch() [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] result = function(*args, **kwargs) [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] return func(*args, **kwargs) [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] raise e [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] nwinfo = self.network_api.allocate_for_instance( [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.497608] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] created_port_ids = self._update_ports_for_instance( [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] with excutils.save_and_reraise_exception(): [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] self.force_reraise() [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] raise self.value [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] updated_port = self._update_port( [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] _ensure_no_port_binding_failure(port) [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.497956] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] raise exception.PortBindingFailed(port_id=port['id']) [ 585.498397] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] nova.exception.PortBindingFailed: Binding failed for port 2a9e5e0b-b590-4ea0-882a-c281c286e7c9, please check neutron logs for more information. [ 585.498397] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] [ 585.498397] env[61857]: INFO nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Terminating instance [ 585.498982] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Acquiring lock "refresh_cache-e7e82254-4088-4232-b861-9aee864c3737" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.636706] env[61857]: DEBUG nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 585.673018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.067s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.673018] env[61857]: ERROR nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0a3cb614-e979-4b79-86df-dbf4a64939ab, please check neutron logs for more information. [ 585.673018] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Traceback (most recent call last): [ 585.673018] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.673018] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] self.driver.spawn(context, instance, image_meta, [ 585.673018] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 585.673018] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.673018] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.673018] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] vm_ref = self.build_virtual_machine(instance, [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] for vif in network_info: [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] return self._sync_wrapper(fn, *args, **kwargs) [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] self.wait() [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] self[:] = self._gt.wait() [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] return self._exit_event.wait() [ 585.673522] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] result = hub.switch() [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] return self.greenlet.switch() [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] result = function(*args, **kwargs) [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] return func(*args, **kwargs) [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] raise e [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] nwinfo = self.network_api.allocate_for_instance( [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.673903] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] created_port_ids = self._update_ports_for_instance( [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] with excutils.save_and_reraise_exception(): [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] self.force_reraise() [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] raise self.value [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] updated_port = self._update_port( [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] _ensure_no_port_binding_failure(port) [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.674327] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] raise exception.PortBindingFailed(port_id=port['id']) [ 585.674672] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] nova.exception.PortBindingFailed: Binding failed for port 0a3cb614-e979-4b79-86df-dbf4a64939ab, please check neutron logs for more information. [ 585.674672] env[61857]: ERROR nova.compute.manager [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] [ 585.674672] env[61857]: DEBUG nova.compute.utils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Binding failed for port 0a3cb614-e979-4b79-86df-dbf4a64939ab, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 585.674672] env[61857]: DEBUG nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Build of instance e8c59da6-c226-4c83-869b-6a200cf1c0b9 was re-scheduled: Binding failed for port 0a3cb614-e979-4b79-86df-dbf4a64939ab, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 585.674672] env[61857]: DEBUG nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 585.674861] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquiring lock "refresh_cache-e8c59da6-c226-4c83-869b-6a200cf1c0b9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.674861] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquired lock "refresh_cache-e8c59da6-c226-4c83-869b-6a200cf1c0b9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.674861] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 585.684149] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 585.684149] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 585.684149] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 585.684317] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 585.684317] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 585.684317] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 585.684405] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 585.684484] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 585.684641] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 585.684797] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 585.684958] env[61857]: DEBUG nova.virt.hardware [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 585.687295] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.481s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.693281] env[61857]: INFO nova.compute.claims [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.700194] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135ed28f-ad62-4290-a9a7-af16feeda021 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.718070] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c13eb78-f42e-4094-a882-a9d6aacfb0e4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.770467] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.818200] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Successfully created port: 19b14803-83fe-4222-9cfa-5921a8bd7b80 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 585.991218] env[61857]: DEBUG oslo_concurrency.lockutils [req-9780af18-208c-43e0-80b0-55e2058f8e78 req-183503ff-c192-4795-b65b-25a9131209c3 service nova] Releasing lock "refresh_cache-e7e82254-4088-4232-b861-9aee864c3737" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.991552] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Acquired lock "refresh_cache-e7e82254-4088-4232-b861-9aee864c3737" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.991730] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 586.273178] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.559564] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.596124] env[61857]: DEBUG nova.compute.manager [req-43b84c20-5259-4759-b47e-f7b832eca521 req-1d6e6cc2-0af8-4784-a14b-4ef4b032f6da service nova] [instance: e7e82254-4088-4232-b861-9aee864c3737] Received event network-vif-deleted-2a9e5e0b-b590-4ea0-882a-c281c286e7c9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 586.777040] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Releasing lock "refresh_cache-e8c59da6-c226-4c83-869b-6a200cf1c0b9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.777040] env[61857]: DEBUG nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 586.777040] env[61857]: DEBUG nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 586.777312] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 586.970186] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.075704] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.239549] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44679bbb-9f2e-4d41-b7fd-395f812e4b37 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.249781] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf821b4-97de-4738-b5b6-20ccbe617b30 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.284177] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a4512d-7859-4215-8980-9a90602627af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.292139] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c0e9a9-722e-4945-be9c-fba7d778db5e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.305795] env[61857]: DEBUG nova.compute.provider_tree [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.475197] env[61857]: DEBUG nova.network.neutron [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.580393] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Releasing lock "refresh_cache-e7e82254-4088-4232-b861-9aee864c3737" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.580393] env[61857]: DEBUG nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 587.580393] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 587.580393] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d085a75-4a1c-4e5a-8f44-b8c0cb652f57 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.590465] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04782d2b-6dbe-4d0c-9e34-4c548152bacb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.617682] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e7e82254-4088-4232-b861-9aee864c3737 could not be found. [ 587.617899] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 587.618200] env[61857]: INFO nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Took 0.04 seconds to destroy the instance on the hypervisor. [ 587.620852] env[61857]: DEBUG oslo.service.loopingcall [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.620852] env[61857]: DEBUG nova.compute.manager [-] [instance: e7e82254-4088-4232-b861-9aee864c3737] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.620852] env[61857]: DEBUG nova.network.neutron [-] [instance: e7e82254-4088-4232-b861-9aee864c3737] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.669160] env[61857]: DEBUG nova.network.neutron [-] [instance: e7e82254-4088-4232-b861-9aee864c3737] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.811077] env[61857]: DEBUG nova.scheduler.client.report [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.939133] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Acquiring lock "952d0a1c-3e94-42dc-9bc4-8063275afc82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.940840] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Lock "952d0a1c-3e94-42dc-9bc4-8063275afc82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.977618] env[61857]: INFO nova.compute.manager [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: e8c59da6-c226-4c83-869b-6a200cf1c0b9] Took 1.20 seconds to deallocate network for instance. [ 587.982105] env[61857]: ERROR nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 19b14803-83fe-4222-9cfa-5921a8bd7b80, please check neutron logs for more information. [ 587.982105] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.982105] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.982105] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.982105] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.982105] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.982105] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.982105] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.982105] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.982105] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 587.982105] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.982105] env[61857]: ERROR nova.compute.manager raise self.value [ 587.982105] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.982105] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.982105] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.982105] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.982742] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.982742] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.982742] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 19b14803-83fe-4222-9cfa-5921a8bd7b80, please check neutron logs for more information. [ 587.982742] env[61857]: ERROR nova.compute.manager [ 587.982742] env[61857]: Traceback (most recent call last): [ 587.982742] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.982742] env[61857]: listener.cb(fileno) [ 587.982742] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.982742] env[61857]: result = function(*args, **kwargs) [ 587.982742] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.982742] env[61857]: return func(*args, **kwargs) [ 587.982742] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.982742] env[61857]: raise e [ 587.982742] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.982742] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 587.982742] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.982742] env[61857]: created_port_ids = self._update_ports_for_instance( [ 587.982742] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.982742] env[61857]: with excutils.save_and_reraise_exception(): [ 587.982742] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.982742] env[61857]: self.force_reraise() [ 587.982742] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.982742] env[61857]: raise self.value [ 587.982742] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.982742] env[61857]: updated_port = self._update_port( [ 587.982742] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.982742] env[61857]: _ensure_no_port_binding_failure(port) [ 587.982742] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.982742] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.983650] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 19b14803-83fe-4222-9cfa-5921a8bd7b80, please check neutron logs for more information. [ 587.983650] env[61857]: Removing descriptor: 16 [ 587.983650] env[61857]: ERROR nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 19b14803-83fe-4222-9cfa-5921a8bd7b80, please check neutron logs for more information. [ 587.983650] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Traceback (most recent call last): [ 587.983650] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 587.983650] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] yield resources [ 587.983650] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.983650] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] self.driver.spawn(context, instance, image_meta, [ 587.983650] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 587.983650] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.983650] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.983650] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] vm_ref = self.build_virtual_machine(instance, [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] for vif in network_info: [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] return self._sync_wrapper(fn, *args, **kwargs) [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] self.wait() [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] self[:] = self._gt.wait() [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] return self._exit_event.wait() [ 587.984053] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] result = hub.switch() [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] return self.greenlet.switch() [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] result = function(*args, **kwargs) [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] return func(*args, **kwargs) [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] raise e [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] nwinfo = self.network_api.allocate_for_instance( [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.984470] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] created_port_ids = self._update_ports_for_instance( [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] with excutils.save_and_reraise_exception(): [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] self.force_reraise() [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] raise self.value [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] updated_port = self._update_port( [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] _ensure_no_port_binding_failure(port) [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.984824] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] raise exception.PortBindingFailed(port_id=port['id']) [ 587.985258] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] nova.exception.PortBindingFailed: Binding failed for port 19b14803-83fe-4222-9cfa-5921a8bd7b80, please check neutron logs for more information. [ 587.985258] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] [ 587.985258] env[61857]: INFO nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Terminating instance [ 587.986677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "refresh_cache-9f079e2f-a22d-48b7-926a-b89ba087d45d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.986771] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquired lock "refresh_cache-9f079e2f-a22d-48b7-926a-b89ba087d45d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.986865] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 588.028643] env[61857]: ERROR nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 715f7af3-9869-4bfa-a356-6e4d4920728d, please check neutron logs for more information. [ 588.028643] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 588.028643] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.028643] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 588.028643] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.028643] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 588.028643] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.028643] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 588.028643] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.028643] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 588.028643] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.028643] env[61857]: ERROR nova.compute.manager raise self.value [ 588.028643] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.028643] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 588.028643] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.028643] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 588.029246] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.029246] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 588.029246] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 715f7af3-9869-4bfa-a356-6e4d4920728d, please check neutron logs for more information. [ 588.029246] env[61857]: ERROR nova.compute.manager [ 588.029246] env[61857]: Traceback (most recent call last): [ 588.029246] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 588.029246] env[61857]: listener.cb(fileno) [ 588.029246] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.029246] env[61857]: result = function(*args, **kwargs) [ 588.029246] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.029246] env[61857]: return func(*args, **kwargs) [ 588.029246] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.029246] env[61857]: raise e [ 588.029246] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.029246] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 588.029246] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.029246] env[61857]: created_port_ids = self._update_ports_for_instance( [ 588.029246] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.029246] env[61857]: with excutils.save_and_reraise_exception(): [ 588.029246] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.029246] env[61857]: self.force_reraise() [ 588.029246] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.029246] env[61857]: raise self.value [ 588.029246] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.029246] env[61857]: updated_port = self._update_port( [ 588.029246] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.029246] env[61857]: _ensure_no_port_binding_failure(port) [ 588.029246] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.029246] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 588.030169] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 715f7af3-9869-4bfa-a356-6e4d4920728d, please check neutron logs for more information. [ 588.030169] env[61857]: Removing descriptor: 17 [ 588.030169] env[61857]: ERROR nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 715f7af3-9869-4bfa-a356-6e4d4920728d, please check neutron logs for more information. [ 588.030169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Traceback (most recent call last): [ 588.030169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 588.030169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] yield resources [ 588.030169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.030169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] self.driver.spawn(context, instance, image_meta, [ 588.030169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 588.030169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.030169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.030169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] vm_ref = self.build_virtual_machine(instance, [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] for vif in network_info: [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] return self._sync_wrapper(fn, *args, **kwargs) [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] self.wait() [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] self[:] = self._gt.wait() [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] return self._exit_event.wait() [ 588.030514] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] result = hub.switch() [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] return self.greenlet.switch() [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] result = function(*args, **kwargs) [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] return func(*args, **kwargs) [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] raise e [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] nwinfo = self.network_api.allocate_for_instance( [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.030899] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] created_port_ids = self._update_ports_for_instance( [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] with excutils.save_and_reraise_exception(): [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] self.force_reraise() [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] raise self.value [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] updated_port = self._update_port( [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] _ensure_no_port_binding_failure(port) [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.031334] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] raise exception.PortBindingFailed(port_id=port['id']) [ 588.031704] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] nova.exception.PortBindingFailed: Binding failed for port 715f7af3-9869-4bfa-a356-6e4d4920728d, please check neutron logs for more information. [ 588.031704] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] [ 588.031704] env[61857]: INFO nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Terminating instance [ 588.032791] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Acquiring lock "refresh_cache-d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.032791] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Acquired lock "refresh_cache-d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.032791] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 588.174727] env[61857]: DEBUG nova.network.neutron [-] [instance: e7e82254-4088-4232-b861-9aee864c3737] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.320120] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.633s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.320872] env[61857]: DEBUG nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 588.327027] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.927s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.328822] env[61857]: INFO nova.compute.claims [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.462626] env[61857]: DEBUG nova.compute.manager [req-905db4d2-7615-40ee-847a-7bc3f464b9b7 req-06a39f6e-54a3-402f-b14f-b47b28a918de service nova] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Received event network-changed-19b14803-83fe-4222-9cfa-5921a8bd7b80 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 588.462851] env[61857]: DEBUG nova.compute.manager [req-905db4d2-7615-40ee-847a-7bc3f464b9b7 req-06a39f6e-54a3-402f-b14f-b47b28a918de service nova] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Refreshing instance network info cache due to event network-changed-19b14803-83fe-4222-9cfa-5921a8bd7b80. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 588.463125] env[61857]: DEBUG oslo_concurrency.lockutils [req-905db4d2-7615-40ee-847a-7bc3f464b9b7 req-06a39f6e-54a3-402f-b14f-b47b28a918de service nova] Acquiring lock "refresh_cache-9f079e2f-a22d-48b7-926a-b89ba087d45d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.520555] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.566183] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.678102] env[61857]: INFO nova.compute.manager [-] [instance: e7e82254-4088-4232-b861-9aee864c3737] Took 1.06 seconds to deallocate network for instance. [ 588.680123] env[61857]: DEBUG nova.compute.claims [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 588.680496] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.705734] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.757343] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.830000] env[61857]: DEBUG nova.compute.utils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.837533] env[61857]: DEBUG nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.837797] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 588.971694] env[61857]: DEBUG nova.policy [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a04613c838145bd9a81a74362e53696', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4487515e1df74f49906bca8c457398b8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 589.024262] env[61857]: INFO nova.scheduler.client.report [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Deleted allocations for instance e8c59da6-c226-4c83-869b-6a200cf1c0b9 [ 589.207686] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Releasing lock "refresh_cache-9f079e2f-a22d-48b7-926a-b89ba087d45d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.208183] env[61857]: DEBUG nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.208401] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 589.208723] env[61857]: DEBUG oslo_concurrency.lockutils [req-905db4d2-7615-40ee-847a-7bc3f464b9b7 req-06a39f6e-54a3-402f-b14f-b47b28a918de service nova] Acquired lock "refresh_cache-9f079e2f-a22d-48b7-926a-b89ba087d45d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.208899] env[61857]: DEBUG nova.network.neutron [req-905db4d2-7615-40ee-847a-7bc3f464b9b7 req-06a39f6e-54a3-402f-b14f-b47b28a918de service nova] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Refreshing network info cache for port 19b14803-83fe-4222-9cfa-5921a8bd7b80 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 589.216175] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68214cd4-1ed6-4d52-8974-bc69aa6d052d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.225629] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcba6e6-69ef-415e-a803-839c5e7f7b6b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.252034] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f079e2f-a22d-48b7-926a-b89ba087d45d could not be found. [ 589.252165] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 589.253616] env[61857]: INFO nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 589.253616] env[61857]: DEBUG oslo.service.loopingcall [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.253616] env[61857]: DEBUG nova.compute.manager [-] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.253616] env[61857]: DEBUG nova.network.neutron [-] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 589.259490] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Releasing lock "refresh_cache-d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.259924] env[61857]: DEBUG nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.260175] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 589.260940] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-980f44cb-7e9a-4122-bb43-743351802d3d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.269189] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041e718c-0e4c-4d16-949a-b8a89e072d09 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.280148] env[61857]: DEBUG nova.network.neutron [-] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.292861] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d4a5719b-6ca1-43ea-8e39-c8026ff44ae6 could not be found. [ 589.293015] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 589.293211] env[61857]: INFO nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 589.293447] env[61857]: DEBUG oslo.service.loopingcall [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.293666] env[61857]: DEBUG nova.compute.manager [-] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.293761] env[61857]: DEBUG nova.network.neutron [-] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 589.322754] env[61857]: DEBUG nova.network.neutron [-] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.335867] env[61857]: DEBUG nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.468257] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Successfully created port: ec73d946-0967-41e1-9a1f-a483dd3a63f9 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 589.535774] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ac8463e-9eaf-41ed-a609-2822b230be32 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "e8c59da6-c226-4c83-869b-6a200cf1c0b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.537s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.608049] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Acquiring lock "768e0e7d-7ebe-4d7a-a971-4fb463731050" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.610019] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Lock "768e0e7d-7ebe-4d7a-a971-4fb463731050" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.613017] env[61857]: DEBUG nova.compute.manager [req-28b06200-121e-4350-8612-1701f94a4f06 req-85011a8c-8238-4e58-be02-9eece041ba4a service nova] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Received event network-changed-715f7af3-9869-4bfa-a356-6e4d4920728d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 589.613017] env[61857]: DEBUG nova.compute.manager [req-28b06200-121e-4350-8612-1701f94a4f06 req-85011a8c-8238-4e58-be02-9eece041ba4a service nova] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Refreshing instance network info cache due to event network-changed-715f7af3-9869-4bfa-a356-6e4d4920728d. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 589.613017] env[61857]: DEBUG oslo_concurrency.lockutils [req-28b06200-121e-4350-8612-1701f94a4f06 req-85011a8c-8238-4e58-be02-9eece041ba4a service nova] Acquiring lock "refresh_cache-d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.613017] env[61857]: DEBUG oslo_concurrency.lockutils [req-28b06200-121e-4350-8612-1701f94a4f06 req-85011a8c-8238-4e58-be02-9eece041ba4a service nova] Acquired lock "refresh_cache-d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.613017] env[61857]: DEBUG nova.network.neutron [req-28b06200-121e-4350-8612-1701f94a4f06 req-85011a8c-8238-4e58-be02-9eece041ba4a service nova] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Refreshing network info cache for port 715f7af3-9869-4bfa-a356-6e4d4920728d {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 589.737477] env[61857]: DEBUG nova.network.neutron [req-905db4d2-7615-40ee-847a-7bc3f464b9b7 req-06a39f6e-54a3-402f-b14f-b47b28a918de service nova] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.783161] env[61857]: DEBUG nova.network.neutron [-] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.824263] env[61857]: DEBUG nova.network.neutron [-] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.828513] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181105be-9e7f-4a2a-a991-564ffe4f66d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.837924] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca5cf53-e896-4387-929c-cd86a7469076 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.845623] env[61857]: DEBUG nova.network.neutron [req-905db4d2-7615-40ee-847a-7bc3f464b9b7 req-06a39f6e-54a3-402f-b14f-b47b28a918de service nova] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.874835] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e5b89b-45d0-4dbb-83c6-306d63a3fedf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.883348] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a1f480-5462-4752-8c96-cda5eb07401c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.898640] env[61857]: DEBUG nova.compute.provider_tree [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.002206] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "b3fd3980-efe4-49a0-b656-32f1543e657c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.003065] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "b3fd3980-efe4-49a0-b656-32f1543e657c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.022605] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Acquiring lock "e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.022844] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Lock "e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.039904] env[61857]: DEBUG nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 590.132296] env[61857]: DEBUG nova.network.neutron [req-28b06200-121e-4350-8612-1701f94a4f06 req-85011a8c-8238-4e58-be02-9eece041ba4a service nova] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.234561] env[61857]: DEBUG nova.network.neutron [req-28b06200-121e-4350-8612-1701f94a4f06 req-85011a8c-8238-4e58-be02-9eece041ba4a service nova] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.287079] env[61857]: INFO nova.compute.manager [-] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Took 1.03 seconds to deallocate network for instance. [ 590.289801] env[61857]: DEBUG nova.compute.claims [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.290603] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.327462] env[61857]: INFO nova.compute.manager [-] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Took 1.03 seconds to deallocate network for instance. [ 590.331029] env[61857]: DEBUG nova.compute.claims [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.331029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.351196] env[61857]: DEBUG nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 590.352988] env[61857]: DEBUG oslo_concurrency.lockutils [req-905db4d2-7615-40ee-847a-7bc3f464b9b7 req-06a39f6e-54a3-402f-b14f-b47b28a918de service nova] Releasing lock "refresh_cache-9f079e2f-a22d-48b7-926a-b89ba087d45d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.373037] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.373037] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.373037] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.373282] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.373282] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.373355] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.373549] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.373704] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.373867] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.374330] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.374651] env[61857]: DEBUG nova.virt.hardware [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.375653] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a3d0a8-a500-42c6-b407-bce94d51fd74 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.386636] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7c140c-40e8-47e0-9392-8f822cb01de1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.403296] env[61857]: DEBUG nova.scheduler.client.report [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.560523] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.619229] env[61857]: DEBUG nova.compute.manager [req-5a934799-470c-4059-ba82-98ebb0300e19 req-08f08633-2e49-4706-a1c7-5f40fe36546e service nova] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Received event network-vif-deleted-19b14803-83fe-4222-9cfa-5921a8bd7b80 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 590.620038] env[61857]: DEBUG nova.compute.manager [req-5a934799-470c-4059-ba82-98ebb0300e19 req-08f08633-2e49-4706-a1c7-5f40fe36546e service nova] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Received event network-changed-ec73d946-0967-41e1-9a1f-a483dd3a63f9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 590.620752] env[61857]: DEBUG nova.compute.manager [req-5a934799-470c-4059-ba82-98ebb0300e19 req-08f08633-2e49-4706-a1c7-5f40fe36546e service nova] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Refreshing instance network info cache due to event network-changed-ec73d946-0967-41e1-9a1f-a483dd3a63f9. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 590.621556] env[61857]: DEBUG oslo_concurrency.lockutils [req-5a934799-470c-4059-ba82-98ebb0300e19 req-08f08633-2e49-4706-a1c7-5f40fe36546e service nova] Acquiring lock "refresh_cache-0ae90dae-d095-418c-abda-f6bb76f89919" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.622183] env[61857]: DEBUG oslo_concurrency.lockutils [req-5a934799-470c-4059-ba82-98ebb0300e19 req-08f08633-2e49-4706-a1c7-5f40fe36546e service nova] Acquired lock "refresh_cache-0ae90dae-d095-418c-abda-f6bb76f89919" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.622183] env[61857]: DEBUG nova.network.neutron [req-5a934799-470c-4059-ba82-98ebb0300e19 req-08f08633-2e49-4706-a1c7-5f40fe36546e service nova] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Refreshing network info cache for port ec73d946-0967-41e1-9a1f-a483dd3a63f9 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 590.670670] env[61857]: ERROR nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ec73d946-0967-41e1-9a1f-a483dd3a63f9, please check neutron logs for more information. [ 590.670670] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 590.670670] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.670670] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 590.670670] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.670670] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 590.670670] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.670670] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 590.670670] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.670670] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 590.670670] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.670670] env[61857]: ERROR nova.compute.manager raise self.value [ 590.670670] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.670670] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 590.670670] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.670670] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 590.671366] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.671366] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 590.671366] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ec73d946-0967-41e1-9a1f-a483dd3a63f9, please check neutron logs for more information. [ 590.671366] env[61857]: ERROR nova.compute.manager [ 590.671366] env[61857]: Traceback (most recent call last): [ 590.671366] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 590.671366] env[61857]: listener.cb(fileno) [ 590.671366] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.671366] env[61857]: result = function(*args, **kwargs) [ 590.671366] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.671366] env[61857]: return func(*args, **kwargs) [ 590.671366] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.671366] env[61857]: raise e [ 590.671366] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.671366] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 590.671366] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.671366] env[61857]: created_port_ids = self._update_ports_for_instance( [ 590.671366] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.671366] env[61857]: with excutils.save_and_reraise_exception(): [ 590.671366] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.671366] env[61857]: self.force_reraise() [ 590.671366] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.671366] env[61857]: raise self.value [ 590.671366] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.671366] env[61857]: updated_port = self._update_port( [ 590.671366] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.671366] env[61857]: _ensure_no_port_binding_failure(port) [ 590.671366] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.671366] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 590.672429] env[61857]: nova.exception.PortBindingFailed: Binding failed for port ec73d946-0967-41e1-9a1f-a483dd3a63f9, please check neutron logs for more information. [ 590.672429] env[61857]: Removing descriptor: 16 [ 590.675242] env[61857]: ERROR nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ec73d946-0967-41e1-9a1f-a483dd3a63f9, please check neutron logs for more information. [ 590.675242] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Traceback (most recent call last): [ 590.675242] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 590.675242] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] yield resources [ 590.675242] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 590.675242] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] self.driver.spawn(context, instance, image_meta, [ 590.675242] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 590.675242] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.675242] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.675242] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] vm_ref = self.build_virtual_machine(instance, [ 590.675242] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] for vif in network_info: [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] return self._sync_wrapper(fn, *args, **kwargs) [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] self.wait() [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] self[:] = self._gt.wait() [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] return self._exit_event.wait() [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 590.675928] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] result = hub.switch() [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] return self.greenlet.switch() [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] result = function(*args, **kwargs) [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] return func(*args, **kwargs) [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] raise e [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] nwinfo = self.network_api.allocate_for_instance( [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] created_port_ids = self._update_ports_for_instance( [ 590.676359] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] with excutils.save_and_reraise_exception(): [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] self.force_reraise() [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] raise self.value [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] updated_port = self._update_port( [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] _ensure_no_port_binding_failure(port) [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] raise exception.PortBindingFailed(port_id=port['id']) [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] nova.exception.PortBindingFailed: Binding failed for port ec73d946-0967-41e1-9a1f-a483dd3a63f9, please check neutron logs for more information. [ 590.676804] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] [ 590.677208] env[61857]: INFO nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Terminating instance [ 590.679649] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Acquiring lock "refresh_cache-0ae90dae-d095-418c-abda-f6bb76f89919" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.739055] env[61857]: DEBUG oslo_concurrency.lockutils [req-28b06200-121e-4350-8612-1701f94a4f06 req-85011a8c-8238-4e58-be02-9eece041ba4a service nova] Releasing lock "refresh_cache-d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.739392] env[61857]: DEBUG nova.compute.manager [req-28b06200-121e-4350-8612-1701f94a4f06 req-85011a8c-8238-4e58-be02-9eece041ba4a service nova] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Received event network-vif-deleted-715f7af3-9869-4bfa-a356-6e4d4920728d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 590.912490] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.585s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.912985] env[61857]: DEBUG nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 590.915582] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquiring lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.915708] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.916178] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.147s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.916397] env[61857]: DEBUG nova.objects.instance [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lazy-loading 'resources' on Instance uuid bb418d62-d47c-42c5-908e-26f1eb034e46 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 591.149677] env[61857]: DEBUG nova.network.neutron [req-5a934799-470c-4059-ba82-98ebb0300e19 req-08f08633-2e49-4706-a1c7-5f40fe36546e service nova] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.184601] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquiring lock "6304714b-8988-4798-990f-d3e19774eb8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.184851] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lock "6304714b-8988-4798-990f-d3e19774eb8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.249229] env[61857]: DEBUG nova.network.neutron [req-5a934799-470c-4059-ba82-98ebb0300e19 req-08f08633-2e49-4706-a1c7-5f40fe36546e service nova] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.418444] env[61857]: DEBUG nova.compute.utils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.419772] env[61857]: DEBUG nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.419943] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 591.480745] env[61857]: DEBUG nova.policy [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0000dbd84e0040d69efd12d16c822e59', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05dc3bb85eda4d1691a954d1b0e366f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 591.752092] env[61857]: DEBUG oslo_concurrency.lockutils [req-5a934799-470c-4059-ba82-98ebb0300e19 req-08f08633-2e49-4706-a1c7-5f40fe36546e service nova] Releasing lock "refresh_cache-0ae90dae-d095-418c-abda-f6bb76f89919" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.752865] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Acquired lock "refresh_cache-0ae90dae-d095-418c-abda-f6bb76f89919" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.752865] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 591.882734] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0264b135-5854-45da-9026-c9dbb5034f59 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.890549] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcf9e4b-ef34-4d5d-abf5-174f214496e4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.923484] env[61857]: DEBUG nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.926820] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6153fe58-f007-4ead-a70e-c29ee678d51c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.931818] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Successfully created port: ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 591.936853] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ea070a-5d4a-48ba-9cc1-0726301b46ed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.950100] env[61857]: DEBUG nova.compute.provider_tree [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.278629] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.350235] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.455264] env[61857]: DEBUG nova.scheduler.client.report [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.669335] env[61857]: DEBUG nova.compute.manager [req-82f263ce-5cc3-4058-b62c-689e65bd69f6 req-3d0edff1-00bf-4bf0-86c8-169273fb6f28 service nova] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Received event network-vif-deleted-ec73d946-0967-41e1-9a1f-a483dd3a63f9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 592.864164] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Releasing lock "refresh_cache-0ae90dae-d095-418c-abda-f6bb76f89919" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.864613] env[61857]: DEBUG nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.864806] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 592.865130] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da9b13cd-b7aa-4573-9b37-6b38a9bc558c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.874872] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bcc3818-59d5-436d-937d-20f9c938343e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.900430] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ae90dae-d095-418c-abda-f6bb76f89919 could not be found. [ 592.900838] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 592.900838] env[61857]: INFO nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Took 0.04 seconds to destroy the instance on the hypervisor. [ 592.901199] env[61857]: DEBUG oslo.service.loopingcall [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.901436] env[61857]: DEBUG nova.compute.manager [-] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.901565] env[61857]: DEBUG nova.network.neutron [-] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 592.927665] env[61857]: DEBUG nova.network.neutron [-] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.940264] env[61857]: DEBUG nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 592.963127] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.046s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.965063] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.613s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.967397] env[61857]: INFO nova.compute.claims [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.974560] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 592.975018] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 592.976019] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 592.976019] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 592.976019] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 592.976019] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 592.976019] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 592.976736] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 592.980028] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 592.980028] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 592.980028] env[61857]: DEBUG nova.virt.hardware [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.980028] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8bd41bb-e37c-4154-ae2e-9f94bc900413 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.988296] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1eacc10-3406-4975-bb9e-50106c66a576 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.992960] env[61857]: INFO nova.scheduler.client.report [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Deleted allocations for instance bb418d62-d47c-42c5-908e-26f1eb034e46 [ 593.136021] env[61857]: ERROR nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf, please check neutron logs for more information. [ 593.136021] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 593.136021] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.136021] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 593.136021] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.136021] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 593.136021] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.136021] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 593.136021] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.136021] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 593.136021] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.136021] env[61857]: ERROR nova.compute.manager raise self.value [ 593.136021] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.136021] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 593.136021] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.136021] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 593.136609] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.136609] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 593.136609] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf, please check neutron logs for more information. [ 593.136609] env[61857]: ERROR nova.compute.manager [ 593.136609] env[61857]: Traceback (most recent call last): [ 593.136609] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 593.136609] env[61857]: listener.cb(fileno) [ 593.136609] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.136609] env[61857]: result = function(*args, **kwargs) [ 593.136609] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.136609] env[61857]: return func(*args, **kwargs) [ 593.136609] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.136609] env[61857]: raise e [ 593.136609] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.136609] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 593.136609] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.136609] env[61857]: created_port_ids = self._update_ports_for_instance( [ 593.136609] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.136609] env[61857]: with excutils.save_and_reraise_exception(): [ 593.136609] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.136609] env[61857]: self.force_reraise() [ 593.136609] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.136609] env[61857]: raise self.value [ 593.136609] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.136609] env[61857]: updated_port = self._update_port( [ 593.136609] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.136609] env[61857]: _ensure_no_port_binding_failure(port) [ 593.136609] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.136609] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 593.137483] env[61857]: nova.exception.PortBindingFailed: Binding failed for port ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf, please check neutron logs for more information. [ 593.137483] env[61857]: Removing descriptor: 16 [ 593.137483] env[61857]: ERROR nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf, please check neutron logs for more information. [ 593.137483] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Traceback (most recent call last): [ 593.137483] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 593.137483] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] yield resources [ 593.137483] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.137483] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] self.driver.spawn(context, instance, image_meta, [ 593.137483] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 593.137483] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.137483] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.137483] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] vm_ref = self.build_virtual_machine(instance, [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] for vif in network_info: [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] return self._sync_wrapper(fn, *args, **kwargs) [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] self.wait() [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] self[:] = self._gt.wait() [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] return self._exit_event.wait() [ 593.137836] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] result = hub.switch() [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] return self.greenlet.switch() [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] result = function(*args, **kwargs) [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] return func(*args, **kwargs) [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] raise e [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] nwinfo = self.network_api.allocate_for_instance( [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.138234] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] created_port_ids = self._update_ports_for_instance( [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] with excutils.save_and_reraise_exception(): [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] self.force_reraise() [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] raise self.value [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] updated_port = self._update_port( [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] _ensure_no_port_binding_failure(port) [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.138606] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] raise exception.PortBindingFailed(port_id=port['id']) [ 593.138957] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] nova.exception.PortBindingFailed: Binding failed for port ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf, please check neutron logs for more information. [ 593.138957] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] [ 593.138957] env[61857]: INFO nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Terminating instance [ 593.141947] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Acquiring lock "refresh_cache-1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.141947] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Acquired lock "refresh_cache-1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.141947] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 593.429920] env[61857]: DEBUG nova.network.neutron [-] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.513395] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b558e40a-9971-40c4-9aa7-7fbf298071db tempest-ServersAdmin275Test-1844277439 tempest-ServersAdmin275Test-1844277439-project-member] Lock "bb418d62-d47c-42c5-908e-26f1eb034e46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.498s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.661755] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.730529] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.935570] env[61857]: INFO nova.compute.manager [-] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Took 1.03 seconds to deallocate network for instance. [ 593.938599] env[61857]: DEBUG nova.compute.claims [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 593.938776] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.235958] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Releasing lock "refresh_cache-1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.235958] env[61857]: DEBUG nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 594.236147] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 594.237411] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d87de72e-de55-4342-b4df-0404d4a5e0cf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.245280] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb6606e-b050-4ba2-a22c-71507bfc1656 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.273816] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6 could not be found. [ 594.273981] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 594.274181] env[61857]: INFO nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 594.274432] env[61857]: DEBUG oslo.service.loopingcall [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 594.274947] env[61857]: DEBUG nova.compute.manager [-] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.274947] env[61857]: DEBUG nova.network.neutron [-] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 594.293810] env[61857]: DEBUG nova.network.neutron [-] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.505237] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf63f395-eee1-4251-8fab-2429ba100e24 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.511367] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16a54d1-ceb6-4295-8294-7e263059595a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.547935] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99141a3f-a6c4-4598-80f4-13f28c52b5c4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.555501] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94c9fbb-cd57-47b8-afac-89d8120cd0d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.571845] env[61857]: DEBUG nova.compute.provider_tree [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.702560] env[61857]: DEBUG nova.compute.manager [req-5fbc143c-eece-4978-8b87-bff44700e1d8 req-e73813b1-1632-47e1-a42e-ce4032caf091 service nova] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Received event network-changed-ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 594.702777] env[61857]: DEBUG nova.compute.manager [req-5fbc143c-eece-4978-8b87-bff44700e1d8 req-e73813b1-1632-47e1-a42e-ce4032caf091 service nova] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Refreshing instance network info cache due to event network-changed-ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 594.702996] env[61857]: DEBUG oslo_concurrency.lockutils [req-5fbc143c-eece-4978-8b87-bff44700e1d8 req-e73813b1-1632-47e1-a42e-ce4032caf091 service nova] Acquiring lock "refresh_cache-1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.703182] env[61857]: DEBUG oslo_concurrency.lockutils [req-5fbc143c-eece-4978-8b87-bff44700e1d8 req-e73813b1-1632-47e1-a42e-ce4032caf091 service nova] Acquired lock "refresh_cache-1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.703319] env[61857]: DEBUG nova.network.neutron [req-5fbc143c-eece-4978-8b87-bff44700e1d8 req-e73813b1-1632-47e1-a42e-ce4032caf091 service nova] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Refreshing network info cache for port ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 594.796860] env[61857]: DEBUG nova.network.neutron [-] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.074324] env[61857]: DEBUG nova.scheduler.client.report [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.299562] env[61857]: INFO nova.compute.manager [-] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Took 1.02 seconds to deallocate network for instance. [ 595.302560] env[61857]: DEBUG nova.compute.claims [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 595.303026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.378434] env[61857]: DEBUG nova.network.neutron [req-5fbc143c-eece-4978-8b87-bff44700e1d8 req-e73813b1-1632-47e1-a42e-ce4032caf091 service nova] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.452532] env[61857]: DEBUG nova.network.neutron [req-5fbc143c-eece-4978-8b87-bff44700e1d8 req-e73813b1-1632-47e1-a42e-ce4032caf091 service nova] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.582969] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.583531] env[61857]: DEBUG nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.591446] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.850s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.955768] env[61857]: DEBUG oslo_concurrency.lockutils [req-5fbc143c-eece-4978-8b87-bff44700e1d8 req-e73813b1-1632-47e1-a42e-ce4032caf091 service nova] Releasing lock "refresh_cache-1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.956113] env[61857]: DEBUG nova.compute.manager [req-5fbc143c-eece-4978-8b87-bff44700e1d8 req-e73813b1-1632-47e1-a42e-ce4032caf091 service nova] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Received event network-vif-deleted-ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 596.090644] env[61857]: DEBUG nova.compute.utils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.096730] env[61857]: DEBUG nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.096878] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 596.166872] env[61857]: DEBUG nova.policy [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df3958a0540d4b999d4e267daa8e392e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f026beafe7464c789744979bb8810adb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.568887] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Successfully created port: e12d32c4-6bf0-422c-b988-cb04aa55820c {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.600025] env[61857]: DEBUG nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.665706] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40504fae-8dba-4412-b94f-2269d3c2fa3c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.674051] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1d9496-a26d-4980-9731-4c3b35c6a5e4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.708451] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75bb6a5-5e90-4198-a98b-138ae32bbc96 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.716140] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4775a10-b1f1-41d1-850e-0f7664552366 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.730816] env[61857]: DEBUG nova.compute.provider_tree [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.894307] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquiring lock "11adaee5-8e56-4679-8373-5d5690a44ca9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.894423] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lock "11adaee5-8e56-4679-8373-5d5690a44ca9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.234370] env[61857]: DEBUG nova.scheduler.client.report [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.516029] env[61857]: DEBUG nova.compute.manager [req-ecf7af13-8c83-4c38-ad2e-b6577d067e0d req-d4beae17-4429-4e54-af5f-fe35e5bfd544 service nova] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Received event network-changed-e12d32c4-6bf0-422c-b988-cb04aa55820c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 597.516254] env[61857]: DEBUG nova.compute.manager [req-ecf7af13-8c83-4c38-ad2e-b6577d067e0d req-d4beae17-4429-4e54-af5f-fe35e5bfd544 service nova] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Refreshing instance network info cache due to event network-changed-e12d32c4-6bf0-422c-b988-cb04aa55820c. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 597.516462] env[61857]: DEBUG oslo_concurrency.lockutils [req-ecf7af13-8c83-4c38-ad2e-b6577d067e0d req-d4beae17-4429-4e54-af5f-fe35e5bfd544 service nova] Acquiring lock "refresh_cache-f14ce7b5-158d-4831-9ba3-404b44752afa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.516605] env[61857]: DEBUG oslo_concurrency.lockutils [req-ecf7af13-8c83-4c38-ad2e-b6577d067e0d req-d4beae17-4429-4e54-af5f-fe35e5bfd544 service nova] Acquired lock "refresh_cache-f14ce7b5-158d-4831-9ba3-404b44752afa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.517240] env[61857]: DEBUG nova.network.neutron [req-ecf7af13-8c83-4c38-ad2e-b6577d067e0d req-d4beae17-4429-4e54-af5f-fe35e5bfd544 service nova] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Refreshing network info cache for port e12d32c4-6bf0-422c-b988-cb04aa55820c {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 597.608936] env[61857]: DEBUG nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.635932] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.636190] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.636347] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.636527] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.636672] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.636821] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.637027] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.637416] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.637416] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.637706] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.637790] env[61857]: DEBUG nova.virt.hardware [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.638660] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2596e0-3e8d-4ae8-9452-d1d78d4142fa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.648556] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aea620d-03b8-43f9-b8bb-6c7d26b6c75e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.741605] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.154s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.742265] env[61857]: ERROR nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 40296f2c-b810-427f-bceb-3b149c9f88e1, please check neutron logs for more information. [ 597.742265] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Traceback (most recent call last): [ 597.742265] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.742265] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] self.driver.spawn(context, instance, image_meta, [ 597.742265] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.742265] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.742265] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.742265] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] vm_ref = self.build_virtual_machine(instance, [ 597.742265] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.742265] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.742265] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] for vif in network_info: [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] return self._sync_wrapper(fn, *args, **kwargs) [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] self.wait() [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] self[:] = self._gt.wait() [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] return self._exit_event.wait() [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] result = hub.switch() [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.742712] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] return self.greenlet.switch() [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] result = function(*args, **kwargs) [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] return func(*args, **kwargs) [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] raise e [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] nwinfo = self.network_api.allocate_for_instance( [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] created_port_ids = self._update_ports_for_instance( [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] with excutils.save_and_reraise_exception(): [ 597.743242] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] self.force_reraise() [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] raise self.value [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] updated_port = self._update_port( [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] _ensure_no_port_binding_failure(port) [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] raise exception.PortBindingFailed(port_id=port['id']) [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] nova.exception.PortBindingFailed: Binding failed for port 40296f2c-b810-427f-bceb-3b149c9f88e1, please check neutron logs for more information. [ 597.743614] env[61857]: ERROR nova.compute.manager [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] [ 597.743954] env[61857]: DEBUG nova.compute.utils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Binding failed for port 40296f2c-b810-427f-bceb-3b149c9f88e1, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.746414] env[61857]: DEBUG nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Build of instance f38e8137-da78-40f7-88a1-0f870a83ffd9 was re-scheduled: Binding failed for port 40296f2c-b810-427f-bceb-3b149c9f88e1, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.748303] env[61857]: DEBUG nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.748544] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Acquiring lock "refresh_cache-f38e8137-da78-40f7-88a1-0f870a83ffd9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.748758] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Acquired lock "refresh_cache-f38e8137-da78-40f7-88a1-0f870a83ffd9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.748856] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.749906] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.050s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.753503] env[61857]: INFO nova.compute.claims [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.906914] env[61857]: ERROR nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e12d32c4-6bf0-422c-b988-cb04aa55820c, please check neutron logs for more information. [ 597.906914] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.906914] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.906914] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.906914] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.906914] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.906914] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.906914] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.906914] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.906914] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 597.906914] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.906914] env[61857]: ERROR nova.compute.manager raise self.value [ 597.906914] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.906914] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.906914] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.906914] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.907519] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.907519] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.907519] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e12d32c4-6bf0-422c-b988-cb04aa55820c, please check neutron logs for more information. [ 597.907519] env[61857]: ERROR nova.compute.manager [ 597.907519] env[61857]: Traceback (most recent call last): [ 597.907519] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.907519] env[61857]: listener.cb(fileno) [ 597.907519] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.907519] env[61857]: result = function(*args, **kwargs) [ 597.907519] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.907519] env[61857]: return func(*args, **kwargs) [ 597.907519] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.907519] env[61857]: raise e [ 597.907519] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.907519] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 597.907519] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.907519] env[61857]: created_port_ids = self._update_ports_for_instance( [ 597.907519] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.907519] env[61857]: with excutils.save_and_reraise_exception(): [ 597.907519] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.907519] env[61857]: self.force_reraise() [ 597.907519] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.907519] env[61857]: raise self.value [ 597.907519] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.907519] env[61857]: updated_port = self._update_port( [ 597.907519] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.907519] env[61857]: _ensure_no_port_binding_failure(port) [ 597.907519] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.907519] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.908509] env[61857]: nova.exception.PortBindingFailed: Binding failed for port e12d32c4-6bf0-422c-b988-cb04aa55820c, please check neutron logs for more information. [ 597.908509] env[61857]: Removing descriptor: 16 [ 597.908509] env[61857]: ERROR nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e12d32c4-6bf0-422c-b988-cb04aa55820c, please check neutron logs for more information. [ 597.908509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Traceback (most recent call last): [ 597.908509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 597.908509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] yield resources [ 597.908509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.908509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] self.driver.spawn(context, instance, image_meta, [ 597.908509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.908509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.908509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.908509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] vm_ref = self.build_virtual_machine(instance, [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] for vif in network_info: [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] return self._sync_wrapper(fn, *args, **kwargs) [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] self.wait() [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] self[:] = self._gt.wait() [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] return self._exit_event.wait() [ 597.908939] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] result = hub.switch() [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] return self.greenlet.switch() [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] result = function(*args, **kwargs) [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] return func(*args, **kwargs) [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] raise e [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] nwinfo = self.network_api.allocate_for_instance( [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.909426] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] created_port_ids = self._update_ports_for_instance( [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] with excutils.save_and_reraise_exception(): [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] self.force_reraise() [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] raise self.value [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] updated_port = self._update_port( [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] _ensure_no_port_binding_failure(port) [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.909898] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] raise exception.PortBindingFailed(port_id=port['id']) [ 597.910423] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] nova.exception.PortBindingFailed: Binding failed for port e12d32c4-6bf0-422c-b988-cb04aa55820c, please check neutron logs for more information. [ 597.910423] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] [ 597.910423] env[61857]: INFO nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Terminating instance [ 597.910423] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "refresh_cache-f14ce7b5-158d-4831-9ba3-404b44752afa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.035055] env[61857]: DEBUG nova.network.neutron [req-ecf7af13-8c83-4c38-ad2e-b6577d067e0d req-d4beae17-4429-4e54-af5f-fe35e5bfd544 service nova] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.137222] env[61857]: DEBUG nova.network.neutron [req-ecf7af13-8c83-4c38-ad2e-b6577d067e0d req-d4beae17-4429-4e54-af5f-fe35e5bfd544 service nova] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.277159] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.411120] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.640103] env[61857]: DEBUG oslo_concurrency.lockutils [req-ecf7af13-8c83-4c38-ad2e-b6577d067e0d req-d4beae17-4429-4e54-af5f-fe35e5bfd544 service nova] Releasing lock "refresh_cache-f14ce7b5-158d-4831-9ba3-404b44752afa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.640532] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquired lock "refresh_cache-f14ce7b5-158d-4831-9ba3-404b44752afa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.640710] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.913812] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Releasing lock "refresh_cache-f38e8137-da78-40f7-88a1-0f870a83ffd9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.914159] env[61857]: DEBUG nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 598.914379] env[61857]: DEBUG nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.914554] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.931869] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.165547] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.229437] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ead17d3-2eb6-4eeb-a606-aba5d4851e30 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.240775] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01651aa2-4db8-4d89-9ce1-fa1633dd5773 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.272244] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f740584-56bc-4d7b-850b-a2af71bbf931 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.280697] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72900fac-6bbe-4cfb-9d7f-b060798143fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.297758] env[61857]: DEBUG nova.compute.provider_tree [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.347336] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.434641] env[61857]: DEBUG nova.network.neutron [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.579190] env[61857]: DEBUG nova.compute.manager [req-21995abf-4c6b-4cf9-ae0b-b6fa6f7a69ba req-30f3eb76-8a95-488e-924f-b3a7aa0cc436 service nova] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Received event network-vif-deleted-e12d32c4-6bf0-422c-b988-cb04aa55820c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 599.800582] env[61857]: DEBUG nova.scheduler.client.report [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.849681] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Releasing lock "refresh_cache-f14ce7b5-158d-4831-9ba3-404b44752afa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.850136] env[61857]: DEBUG nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.850800] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 599.850800] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-077643cf-cd05-4655-8317-03b4fdbb98fe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.859482] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6223741e-f692-4eb8-b1e9-9063829fb112 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.882298] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f14ce7b5-158d-4831-9ba3-404b44752afa could not be found. [ 599.882298] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 599.882298] env[61857]: INFO nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Took 0.03 seconds to destroy the instance on the hypervisor. [ 599.882458] env[61857]: DEBUG oslo.service.loopingcall [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.882891] env[61857]: DEBUG nova.compute.manager [-] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.882986] env[61857]: DEBUG nova.network.neutron [-] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.902696] env[61857]: DEBUG nova.network.neutron [-] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.937825] env[61857]: INFO nova.compute.manager [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] [instance: f38e8137-da78-40f7-88a1-0f870a83ffd9] Took 1.02 seconds to deallocate network for instance. [ 600.309065] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.309908] env[61857]: DEBUG nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.314382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.261s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.315182] env[61857]: DEBUG nova.objects.instance [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lazy-loading 'resources' on Instance uuid 2e7c154b-dab2-4384-bdb8-d098967654c5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 600.408336] env[61857]: DEBUG nova.network.neutron [-] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.824376] env[61857]: DEBUG nova.compute.utils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.827238] env[61857]: DEBUG nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.827577] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 600.887340] env[61857]: DEBUG nova.policy [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c0bffed5b744365b4c7119d1d9950de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c9b86bc2de140c2a7381bc862c9b329', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.914639] env[61857]: INFO nova.compute.manager [-] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Took 1.03 seconds to deallocate network for instance. [ 600.917097] env[61857]: DEBUG nova.compute.claims [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.917258] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.966203] env[61857]: INFO nova.scheduler.client.report [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Deleted allocations for instance f38e8137-da78-40f7-88a1-0f870a83ffd9 [ 601.317990] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6bca628-b5c1-4987-93dd-2acd00106e16 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.326084] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e48b63-412f-458d-a433-85c217fe0fa4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.332123] env[61857]: DEBUG nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.374125] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992c1ca9-604e-4494-8991-36d5e0dfd177 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.383697] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ab20cf-b61e-47bb-96bd-119ca9276ea7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.399053] env[61857]: DEBUG nova.compute.provider_tree [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.426000] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Successfully created port: 5551e212-0280-4973-8405-b66be8d415ae {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.485457] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed8e37a4-5e6f-4e78-8014-31d0e1960c00 tempest-InstanceActionsNegativeTestJSON-293583255 tempest-InstanceActionsNegativeTestJSON-293583255-project-member] Lock "f38e8137-da78-40f7-88a1-0f870a83ffd9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.135s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.904021] env[61857]: DEBUG nova.scheduler.client.report [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.991661] env[61857]: DEBUG nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 602.348261] env[61857]: DEBUG nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.378107] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.378107] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.378318] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.378886] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.378886] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.378886] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.379055] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.379237] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.379445] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.379880] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.380159] env[61857]: DEBUG nova.virt.hardware [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.381162] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04d494e-5354-4965-b51f-7d4a6186af62 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.392751] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f69e83-9948-4550-9da9-3b9b979b435e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.410416] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.096s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.412969] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.760s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.441531] env[61857]: INFO nova.scheduler.client.report [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Deleted allocations for instance 2e7c154b-dab2-4384-bdb8-d098967654c5 [ 602.516789] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.770278] env[61857]: ERROR nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5551e212-0280-4973-8405-b66be8d415ae, please check neutron logs for more information. [ 602.770278] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.770278] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.770278] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.770278] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.770278] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.770278] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.770278] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.770278] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.770278] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 602.770278] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.770278] env[61857]: ERROR nova.compute.manager raise self.value [ 602.770278] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.770278] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.770278] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.770278] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.771734] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.771734] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.771734] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5551e212-0280-4973-8405-b66be8d415ae, please check neutron logs for more information. [ 602.771734] env[61857]: ERROR nova.compute.manager [ 602.771734] env[61857]: Traceback (most recent call last): [ 602.771734] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.771734] env[61857]: listener.cb(fileno) [ 602.771734] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.771734] env[61857]: result = function(*args, **kwargs) [ 602.771734] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.771734] env[61857]: return func(*args, **kwargs) [ 602.771734] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.771734] env[61857]: raise e [ 602.771734] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.771734] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 602.771734] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.771734] env[61857]: created_port_ids = self._update_ports_for_instance( [ 602.771734] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.771734] env[61857]: with excutils.save_and_reraise_exception(): [ 602.771734] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.771734] env[61857]: self.force_reraise() [ 602.771734] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.771734] env[61857]: raise self.value [ 602.771734] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.771734] env[61857]: updated_port = self._update_port( [ 602.771734] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.771734] env[61857]: _ensure_no_port_binding_failure(port) [ 602.771734] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.771734] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.774359] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 5551e212-0280-4973-8405-b66be8d415ae, please check neutron logs for more information. [ 602.774359] env[61857]: Removing descriptor: 16 [ 602.774359] env[61857]: ERROR nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5551e212-0280-4973-8405-b66be8d415ae, please check neutron logs for more information. [ 602.774359] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Traceback (most recent call last): [ 602.774359] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 602.774359] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] yield resources [ 602.774359] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.774359] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] self.driver.spawn(context, instance, image_meta, [ 602.774359] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.774359] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.774359] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.774359] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] vm_ref = self.build_virtual_machine(instance, [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] for vif in network_info: [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] return self._sync_wrapper(fn, *args, **kwargs) [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] self.wait() [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] self[:] = self._gt.wait() [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] return self._exit_event.wait() [ 602.775196] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] result = hub.switch() [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] return self.greenlet.switch() [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] result = function(*args, **kwargs) [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] return func(*args, **kwargs) [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] raise e [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] nwinfo = self.network_api.allocate_for_instance( [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.776048] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] created_port_ids = self._update_ports_for_instance( [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] with excutils.save_and_reraise_exception(): [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] self.force_reraise() [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] raise self.value [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] updated_port = self._update_port( [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] _ensure_no_port_binding_failure(port) [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.776733] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] raise exception.PortBindingFailed(port_id=port['id']) [ 602.777198] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] nova.exception.PortBindingFailed: Binding failed for port 5551e212-0280-4973-8405-b66be8d415ae, please check neutron logs for more information. [ 602.777198] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] [ 602.777198] env[61857]: INFO nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Terminating instance [ 602.777198] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Acquiring lock "refresh_cache-5ad7fc93-72b3-4108-923d-f52d74589423" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.777198] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Acquired lock "refresh_cache-5ad7fc93-72b3-4108-923d-f52d74589423" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.777198] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 602.843861] env[61857]: DEBUG nova.compute.manager [req-e77c497f-e711-41b9-8305-dd20a43e6a90 req-2f0d4e4e-6a43-490e-bc11-99e981ca8092 service nova] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Received event network-changed-5551e212-0280-4973-8405-b66be8d415ae {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 602.844075] env[61857]: DEBUG nova.compute.manager [req-e77c497f-e711-41b9-8305-dd20a43e6a90 req-2f0d4e4e-6a43-490e-bc11-99e981ca8092 service nova] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Refreshing instance network info cache due to event network-changed-5551e212-0280-4973-8405-b66be8d415ae. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 602.844267] env[61857]: DEBUG oslo_concurrency.lockutils [req-e77c497f-e711-41b9-8305-dd20a43e6a90 req-2f0d4e4e-6a43-490e-bc11-99e981ca8092 service nova] Acquiring lock "refresh_cache-5ad7fc93-72b3-4108-923d-f52d74589423" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.953047] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9abf99e-2713-48c2-8ac8-06dc51ba5b38 tempest-ServerDiagnosticsV248Test-1392553915 tempest-ServerDiagnosticsV248Test-1392553915-project-member] Lock "2e7c154b-dab2-4384-bdb8-d098967654c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.798s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.300959] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.361119] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e91de35-6689-4bef-aa9e-61f122a114ea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.370973] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562a511d-e08d-49e2-a4fc-ecb469841ea8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.402455] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51820a2f-622a-49e0-81ab-d9dd6da59615 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.410671] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d7fb71-50f0-4fb0-95ef-7df7c97573d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.425442] env[61857]: DEBUG nova.compute.provider_tree [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.427395] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.930988] env[61857]: DEBUG nova.scheduler.client.report [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.933770] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Releasing lock "refresh_cache-5ad7fc93-72b3-4108-923d-f52d74589423" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.933933] env[61857]: DEBUG nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.934149] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 603.934642] env[61857]: DEBUG oslo_concurrency.lockutils [req-e77c497f-e711-41b9-8305-dd20a43e6a90 req-2f0d4e4e-6a43-490e-bc11-99e981ca8092 service nova] Acquired lock "refresh_cache-5ad7fc93-72b3-4108-923d-f52d74589423" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.938021] env[61857]: DEBUG nova.network.neutron [req-e77c497f-e711-41b9-8305-dd20a43e6a90 req-2f0d4e4e-6a43-490e-bc11-99e981ca8092 service nova] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Refreshing network info cache for port 5551e212-0280-4973-8405-b66be8d415ae {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 603.938021] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9fe868eb-df54-49dc-8d28-989bf9b65666 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.945272] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130348ff-ac9b-47d0-b824-c2207f38a3a8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.971726] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ad7fc93-72b3-4108-923d-f52d74589423 could not be found. [ 603.972067] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 603.972453] env[61857]: INFO nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Took 0.04 seconds to destroy the instance on the hypervisor. [ 603.972749] env[61857]: DEBUG oslo.service.loopingcall [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.973153] env[61857]: DEBUG nova.compute.manager [-] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.973308] env[61857]: DEBUG nova.network.neutron [-] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 603.993900] env[61857]: DEBUG nova.network.neutron [-] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.437012] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.024s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.437690] env[61857]: ERROR nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 53eb146e-3d62-4715-b230-6f04afc7b67c, please check neutron logs for more information. [ 604.437690] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Traceback (most recent call last): [ 604.437690] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.437690] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] self.driver.spawn(context, instance, image_meta, [ 604.437690] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 604.437690] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.437690] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.437690] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] vm_ref = self.build_virtual_machine(instance, [ 604.437690] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.437690] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.437690] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] for vif in network_info: [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] return self._sync_wrapper(fn, *args, **kwargs) [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] self.wait() [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] self[:] = self._gt.wait() [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] return self._exit_event.wait() [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] result = hub.switch() [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.438080] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] return self.greenlet.switch() [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] result = function(*args, **kwargs) [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] return func(*args, **kwargs) [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] raise e [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] nwinfo = self.network_api.allocate_for_instance( [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] created_port_ids = self._update_ports_for_instance( [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] with excutils.save_and_reraise_exception(): [ 604.438456] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] self.force_reraise() [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] raise self.value [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] updated_port = self._update_port( [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] _ensure_no_port_binding_failure(port) [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] raise exception.PortBindingFailed(port_id=port['id']) [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] nova.exception.PortBindingFailed: Binding failed for port 53eb146e-3d62-4715-b230-6f04afc7b67c, please check neutron logs for more information. [ 604.438956] env[61857]: ERROR nova.compute.manager [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] [ 604.439298] env[61857]: DEBUG nova.compute.utils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Binding failed for port 53eb146e-3d62-4715-b230-6f04afc7b67c, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.441519] env[61857]: DEBUG nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Build of instance 0b6029b8-3267-4662-a1af-8e0b6af3288f was re-scheduled: Binding failed for port 53eb146e-3d62-4715-b230-6f04afc7b67c, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.441994] env[61857]: DEBUG nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.442215] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquiring lock "refresh_cache-0b6029b8-3267-4662-a1af-8e0b6af3288f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.442566] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Acquired lock "refresh_cache-0b6029b8-3267-4662-a1af-8e0b6af3288f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.442566] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 604.443532] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.763s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.484143] env[61857]: DEBUG nova.network.neutron [req-e77c497f-e711-41b9-8305-dd20a43e6a90 req-2f0d4e4e-6a43-490e-bc11-99e981ca8092 service nova] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.496607] env[61857]: DEBUG nova.network.neutron [-] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.698917] env[61857]: DEBUG nova.network.neutron [req-e77c497f-e711-41b9-8305-dd20a43e6a90 req-2f0d4e4e-6a43-490e-bc11-99e981ca8092 service nova] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.977389] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.001972] env[61857]: INFO nova.compute.manager [-] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Took 1.03 seconds to deallocate network for instance. [ 605.005231] env[61857]: DEBUG nova.compute.claims [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 605.006365] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.142304] env[61857]: DEBUG nova.compute.manager [req-36bcfb93-5110-48ff-9a34-0cac936af84d req-41fa9840-fb36-4f5b-8e26-cdf9d86ec3c6 service nova] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Received event network-vif-deleted-5551e212-0280-4973-8405-b66be8d415ae {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 605.151790] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.203519] env[61857]: DEBUG oslo_concurrency.lockutils [req-e77c497f-e711-41b9-8305-dd20a43e6a90 req-2f0d4e4e-6a43-490e-bc11-99e981ca8092 service nova] Releasing lock "refresh_cache-5ad7fc93-72b3-4108-923d-f52d74589423" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.289770] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 605.289842] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 605.416706] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bcc6bf-631a-4d89-87d6-014fecdc6d55 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.424733] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929b2352-6424-420c-a564-74571c2f0b69 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.457322] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bf7ab1-777c-4d41-a5ba-9b6c7b97b6f8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.465737] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606a4216-5bb6-43de-9cf3-46a8fa540e1a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.481501] env[61857]: DEBUG nova.compute.provider_tree [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.655426] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Releasing lock "refresh_cache-0b6029b8-3267-4662-a1af-8e0b6af3288f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.655677] env[61857]: DEBUG nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.656197] env[61857]: DEBUG nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.656197] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 605.675784] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.805375] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 605.805554] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 605.805687] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Rebuilding the list of instances to heal {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 605.983479] env[61857]: DEBUG nova.scheduler.client.report [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.178282] env[61857]: DEBUG nova.network.neutron [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.310664] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 606.310833] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 606.311044] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 606.311136] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 606.311224] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 606.311347] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 606.311472] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Didn't find any instances for network info cache update. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 606.311667] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.311886] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.311953] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.312112] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.312258] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.312426] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.312562] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 606.312712] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.488704] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.045s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.489459] env[61857]: ERROR nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2a9e5e0b-b590-4ea0-882a-c281c286e7c9, please check neutron logs for more information. [ 606.489459] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] Traceback (most recent call last): [ 606.489459] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.489459] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] self.driver.spawn(context, instance, image_meta, [ 606.489459] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 606.489459] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.489459] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.489459] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] vm_ref = self.build_virtual_machine(instance, [ 606.489459] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.489459] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.489459] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] for vif in network_info: [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] return self._sync_wrapper(fn, *args, **kwargs) [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] self.wait() [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] self[:] = self._gt.wait() [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] return self._exit_event.wait() [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] result = hub.switch() [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.489864] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] return self.greenlet.switch() [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] result = function(*args, **kwargs) [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] return func(*args, **kwargs) [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] raise e [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] nwinfo = self.network_api.allocate_for_instance( [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] created_port_ids = self._update_ports_for_instance( [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] with excutils.save_and_reraise_exception(): [ 606.490520] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] self.force_reraise() [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] raise self.value [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] updated_port = self._update_port( [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] _ensure_no_port_binding_failure(port) [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] raise exception.PortBindingFailed(port_id=port['id']) [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] nova.exception.PortBindingFailed: Binding failed for port 2a9e5e0b-b590-4ea0-882a-c281c286e7c9, please check neutron logs for more information. [ 606.491064] env[61857]: ERROR nova.compute.manager [instance: e7e82254-4088-4232-b861-9aee864c3737] [ 606.491455] env[61857]: DEBUG nova.compute.utils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Binding failed for port 2a9e5e0b-b590-4ea0-882a-c281c286e7c9, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 606.491924] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.202s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.499362] env[61857]: DEBUG nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Build of instance e7e82254-4088-4232-b861-9aee864c3737 was re-scheduled: Binding failed for port 2a9e5e0b-b590-4ea0-882a-c281c286e7c9, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 606.499362] env[61857]: DEBUG nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 606.499362] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Acquiring lock "refresh_cache-e7e82254-4088-4232-b861-9aee864c3737" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.499362] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Acquired lock "refresh_cache-e7e82254-4088-4232-b861-9aee864c3737" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.499632] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 606.681756] env[61857]: INFO nova.compute.manager [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] [instance: 0b6029b8-3267-4662-a1af-8e0b6af3288f] Took 1.03 seconds to deallocate network for instance. [ 606.816479] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.031930] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.204152] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.523737] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85d7492-d11b-403e-8799-b3b34b22bd3d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.534156] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce0008d-ab94-42b4-9652-6e326c1bc141 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.569401] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f65088-2e11-4797-a815-3963e939c8d3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.576978] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb79e49-876d-4576-89ef-a35da6b8753a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.590888] env[61857]: DEBUG nova.compute.provider_tree [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.708776] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Releasing lock "refresh_cache-e7e82254-4088-4232-b861-9aee864c3737" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.708776] env[61857]: DEBUG nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 607.708776] env[61857]: DEBUG nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.709039] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 607.713871] env[61857]: INFO nova.scheduler.client.report [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Deleted allocations for instance 0b6029b8-3267-4662-a1af-8e0b6af3288f [ 607.740116] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.094775] env[61857]: DEBUG nova.scheduler.client.report [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.223999] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d4c911e0-f464-4071-b23a-0926875fce62 tempest-ServersAdminTestJSON-1223101990 tempest-ServersAdminTestJSON-1223101990-project-member] Lock "0b6029b8-3267-4662-a1af-8e0b6af3288f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.002s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.243214] env[61857]: DEBUG nova.network.neutron [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.623121] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.113s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.623121] env[61857]: ERROR nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 19b14803-83fe-4222-9cfa-5921a8bd7b80, please check neutron logs for more information. [ 608.623121] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Traceback (most recent call last): [ 608.623121] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.623121] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] self.driver.spawn(context, instance, image_meta, [ 608.623121] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 608.623121] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.623121] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.623121] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] vm_ref = self.build_virtual_machine(instance, [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] for vif in network_info: [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] return self._sync_wrapper(fn, *args, **kwargs) [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] self.wait() [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] self[:] = self._gt.wait() [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] return self._exit_event.wait() [ 608.623718] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] result = hub.switch() [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] return self.greenlet.switch() [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] result = function(*args, **kwargs) [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] return func(*args, **kwargs) [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] raise e [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] nwinfo = self.network_api.allocate_for_instance( [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.624241] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] created_port_ids = self._update_ports_for_instance( [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] with excutils.save_and_reraise_exception(): [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] self.force_reraise() [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] raise self.value [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] updated_port = self._update_port( [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] _ensure_no_port_binding_failure(port) [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.624874] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] raise exception.PortBindingFailed(port_id=port['id']) [ 608.625505] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] nova.exception.PortBindingFailed: Binding failed for port 19b14803-83fe-4222-9cfa-5921a8bd7b80, please check neutron logs for more information. [ 608.625505] env[61857]: ERROR nova.compute.manager [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] [ 608.625505] env[61857]: DEBUG nova.compute.utils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Binding failed for port 19b14803-83fe-4222-9cfa-5921a8bd7b80, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 608.625505] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.280s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.625505] env[61857]: DEBUG nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Build of instance 9f079e2f-a22d-48b7-926a-b89ba087d45d was re-scheduled: Binding failed for port 19b14803-83fe-4222-9cfa-5921a8bd7b80, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 608.625677] env[61857]: DEBUG nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 608.625677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "refresh_cache-9f079e2f-a22d-48b7-926a-b89ba087d45d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.625677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquired lock "refresh_cache-9f079e2f-a22d-48b7-926a-b89ba087d45d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.625677] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 608.732935] env[61857]: DEBUG nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.752304] env[61857]: INFO nova.compute.manager [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] [instance: e7e82254-4088-4232-b861-9aee864c3737] Took 1.04 seconds to deallocate network for instance. [ 609.159980] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.262623] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.504411] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.558064] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7017ca5d-0ed5-4711-b0da-145dd9ac51a8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.567017] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032db9ef-3172-43b5-90fb-c8ff9e7cf1b0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.600586] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f91bc99-9565-4cb1-8f5f-0cc2d2c5eb2e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.609875] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bd6bff-f24c-4bbb-b7dd-6ab45cc4cde4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.624657] env[61857]: DEBUG nova.compute.provider_tree [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.792616] env[61857]: INFO nova.scheduler.client.report [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Deleted allocations for instance e7e82254-4088-4232-b861-9aee864c3737 [ 610.007815] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Releasing lock "refresh_cache-9f079e2f-a22d-48b7-926a-b89ba087d45d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.008127] env[61857]: DEBUG nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 610.008500] env[61857]: DEBUG nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.008561] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.036122] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.129313] env[61857]: DEBUG nova.scheduler.client.report [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.304182] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8def678e-e403-44ca-9d96-df1cad2631eb tempest-ServersWithSpecificFlavorTestJSON-1313636939 tempest-ServersWithSpecificFlavorTestJSON-1313636939-project-member] Lock "e7e82254-4088-4232-b861-9aee864c3737" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.848s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.544018] env[61857]: DEBUG nova.network.neutron [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.639893] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.029s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.639893] env[61857]: ERROR nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 715f7af3-9869-4bfa-a356-6e4d4920728d, please check neutron logs for more information. [ 610.639893] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Traceback (most recent call last): [ 610.639893] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.639893] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] self.driver.spawn(context, instance, image_meta, [ 610.639893] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 610.639893] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.639893] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.639893] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] vm_ref = self.build_virtual_machine(instance, [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] for vif in network_info: [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] return self._sync_wrapper(fn, *args, **kwargs) [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] self.wait() [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] self[:] = self._gt.wait() [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] return self._exit_event.wait() [ 610.640186] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] result = hub.switch() [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] return self.greenlet.switch() [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] result = function(*args, **kwargs) [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] return func(*args, **kwargs) [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] raise e [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] nwinfo = self.network_api.allocate_for_instance( [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.640615] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] created_port_ids = self._update_ports_for_instance( [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] with excutils.save_and_reraise_exception(): [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] self.force_reraise() [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] raise self.value [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] updated_port = self._update_port( [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] _ensure_no_port_binding_failure(port) [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.640871] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] raise exception.PortBindingFailed(port_id=port['id']) [ 610.641169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] nova.exception.PortBindingFailed: Binding failed for port 715f7af3-9869-4bfa-a356-6e4d4920728d, please check neutron logs for more information. [ 610.641169] env[61857]: ERROR nova.compute.manager [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] [ 610.647022] env[61857]: DEBUG nova.compute.utils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Binding failed for port 715f7af3-9869-4bfa-a356-6e4d4920728d, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 610.647022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.083s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.647022] env[61857]: INFO nova.compute.claims [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 610.648219] env[61857]: DEBUG nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Build of instance d4a5719b-6ca1-43ea-8e39-c8026ff44ae6 was re-scheduled: Binding failed for port 715f7af3-9869-4bfa-a356-6e4d4920728d, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 610.648839] env[61857]: DEBUG nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 610.650288] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Acquiring lock "refresh_cache-d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.650288] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Acquired lock "refresh_cache-d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.650671] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 610.807019] env[61857]: DEBUG nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 611.050216] env[61857]: INFO nova.compute.manager [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: 9f079e2f-a22d-48b7-926a-b89ba087d45d] Took 1.04 seconds to deallocate network for instance. [ 611.193469] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.341389] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.394190] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.900855] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Releasing lock "refresh_cache-d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.901135] env[61857]: DEBUG nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 611.901358] env[61857]: DEBUG nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.901509] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.921420] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.091889] env[61857]: INFO nova.scheduler.client.report [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Deleted allocations for instance 9f079e2f-a22d-48b7-926a-b89ba087d45d [ 612.196038] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553ab685-c812-4653-a886-c4329e708500 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.203407] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fd985c-7f4d-44ad-b039-5031522446e7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.237988] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8b38ef-172b-4495-ae6f-0baca256b3ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.246376] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e54e8727-60a0-423b-af3c-bdabf17da63f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.266027] env[61857]: DEBUG nova.compute.provider_tree [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.424832] env[61857]: DEBUG nova.network.neutron [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.601304] env[61857]: DEBUG oslo_concurrency.lockutils [None req-91d067f4-ed9c-41c3-a9e5-4952a38ad712 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "9f079e2f-a22d-48b7-926a-b89ba087d45d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.372s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.763900] env[61857]: DEBUG nova.scheduler.client.report [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.927602] env[61857]: INFO nova.compute.manager [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] [instance: d4a5719b-6ca1-43ea-8e39-c8026ff44ae6] Took 1.03 seconds to deallocate network for instance. [ 613.109394] env[61857]: DEBUG nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 613.270241] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.626s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.272043] env[61857]: DEBUG nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 613.279892] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.341s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.630625] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.781727] env[61857]: DEBUG nova.compute.utils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 613.783175] env[61857]: DEBUG nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 613.787281] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 613.893261] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "e88aed62-6561-4e29-b041-46b19cc5ef63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.893499] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "e88aed62-6561-4e29-b041-46b19cc5ef63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.905840] env[61857]: DEBUG nova.policy [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df3958a0540d4b999d4e267daa8e392e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f026beafe7464c789744979bb8810adb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 613.975414] env[61857]: INFO nova.scheduler.client.report [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Deleted allocations for instance d4a5719b-6ca1-43ea-8e39-c8026ff44ae6 [ 614.257672] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fa1269-9c38-4434-a27b-4421c40cc866 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.264122] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6961b6b5-fa70-400f-a15d-6abd83d64635 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.299522] env[61857]: DEBUG nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 614.302896] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ca6c3c-71b5-4f3f-9797-064c23b948c8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.311704] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589e8aef-888d-421d-ac09-d4b0bca2c124 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.329864] env[61857]: DEBUG nova.compute.provider_tree [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.490284] env[61857]: DEBUG oslo_concurrency.lockutils [None req-46fc5599-c701-4f81-84ee-80b000ee1803 tempest-VolumesAssistedSnapshotsTest-1783112840 tempest-VolumesAssistedSnapshotsTest-1783112840-project-member] Lock "d4a5719b-6ca1-43ea-8e39-c8026ff44ae6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.024s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.695919] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Successfully created port: ea4ced50-70d2-45b3-b485-75acc9a5a73b {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 614.831285] env[61857]: DEBUG nova.scheduler.client.report [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.993649] env[61857]: DEBUG nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.103022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "c4f70321-f612-416e-b209-eb974dab9d49" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.104086] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "c4f70321-f612-416e-b209-eb974dab9d49" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.317195] env[61857]: DEBUG nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 615.338606] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.057s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.338606] env[61857]: ERROR nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ec73d946-0967-41e1-9a1f-a483dd3a63f9, please check neutron logs for more information. [ 615.338606] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Traceback (most recent call last): [ 615.338606] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.338606] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] self.driver.spawn(context, instance, image_meta, [ 615.338606] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 615.338606] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.338606] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.338606] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] vm_ref = self.build_virtual_machine(instance, [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] for vif in network_info: [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] return self._sync_wrapper(fn, *args, **kwargs) [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] self.wait() [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] self[:] = self._gt.wait() [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] return self._exit_event.wait() [ 615.338864] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] result = hub.switch() [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] return self.greenlet.switch() [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] result = function(*args, **kwargs) [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] return func(*args, **kwargs) [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] raise e [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] nwinfo = self.network_api.allocate_for_instance( [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.339198] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] created_port_ids = self._update_ports_for_instance( [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] with excutils.save_and_reraise_exception(): [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] self.force_reraise() [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] raise self.value [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] updated_port = self._update_port( [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] _ensure_no_port_binding_failure(port) [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.339506] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] raise exception.PortBindingFailed(port_id=port['id']) [ 615.339799] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] nova.exception.PortBindingFailed: Binding failed for port ec73d946-0967-41e1-9a1f-a483dd3a63f9, please check neutron logs for more information. [ 615.339799] env[61857]: ERROR nova.compute.manager [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] [ 615.342916] env[61857]: DEBUG nova.compute.utils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Binding failed for port ec73d946-0967-41e1-9a1f-a483dd3a63f9, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 615.346108] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.043s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.351832] env[61857]: DEBUG nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Build of instance 0ae90dae-d095-418c-abda-f6bb76f89919 was re-scheduled: Binding failed for port ec73d946-0967-41e1-9a1f-a483dd3a63f9, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 615.352234] env[61857]: DEBUG nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 615.353021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Acquiring lock "refresh_cache-0ae90dae-d095-418c-abda-f6bb76f89919" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.353021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Acquired lock "refresh_cache-0ae90dae-d095-418c-abda-f6bb76f89919" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.353021] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.361434] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 615.361694] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 615.362080] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 615.362478] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 615.362652] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 615.362814] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 615.363476] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 615.363476] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 615.363713] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 615.363713] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 615.364299] env[61857]: DEBUG nova.virt.hardware [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 615.365245] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563f60e2-29aa-486b-a743-3066f8637e29 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.381697] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9a7cc2-56e8-4fc9-92e4-76e0b5a3b29e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.519905] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.888946] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.998081] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.190659] env[61857]: DEBUG nova.compute.manager [req-d7c0e17b-e460-4e85-8e97-4d539cb12687 req-5a046a2a-7ebd-435e-8224-ae4de5d78b35 service nova] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Received event network-changed-ea4ced50-70d2-45b3-b485-75acc9a5a73b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 616.190896] env[61857]: DEBUG nova.compute.manager [req-d7c0e17b-e460-4e85-8e97-4d539cb12687 req-5a046a2a-7ebd-435e-8224-ae4de5d78b35 service nova] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Refreshing instance network info cache due to event network-changed-ea4ced50-70d2-45b3-b485-75acc9a5a73b. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 616.191481] env[61857]: DEBUG oslo_concurrency.lockutils [req-d7c0e17b-e460-4e85-8e97-4d539cb12687 req-5a046a2a-7ebd-435e-8224-ae4de5d78b35 service nova] Acquiring lock "refresh_cache-b54af0ef-448e-453c-9056-32ad9141d9d3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.191676] env[61857]: DEBUG oslo_concurrency.lockutils [req-d7c0e17b-e460-4e85-8e97-4d539cb12687 req-5a046a2a-7ebd-435e-8224-ae4de5d78b35 service nova] Acquired lock "refresh_cache-b54af0ef-448e-453c-9056-32ad9141d9d3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.191903] env[61857]: DEBUG nova.network.neutron [req-d7c0e17b-e460-4e85-8e97-4d539cb12687 req-5a046a2a-7ebd-435e-8224-ae4de5d78b35 service nova] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Refreshing network info cache for port ea4ced50-70d2-45b3-b485-75acc9a5a73b {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 616.364821] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6fa041-8b9e-483c-9052-0a21fc987802 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.369489] env[61857]: ERROR nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ea4ced50-70d2-45b3-b485-75acc9a5a73b, please check neutron logs for more information. [ 616.369489] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 616.369489] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.369489] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 616.369489] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.369489] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 616.369489] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.369489] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 616.369489] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.369489] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 616.369489] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.369489] env[61857]: ERROR nova.compute.manager raise self.value [ 616.369489] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.369489] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 616.369489] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.369489] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 616.369919] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.369919] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 616.369919] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ea4ced50-70d2-45b3-b485-75acc9a5a73b, please check neutron logs for more information. [ 616.369919] env[61857]: ERROR nova.compute.manager [ 616.369919] env[61857]: Traceback (most recent call last): [ 616.369919] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 616.369919] env[61857]: listener.cb(fileno) [ 616.369919] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.369919] env[61857]: result = function(*args, **kwargs) [ 616.369919] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.369919] env[61857]: return func(*args, **kwargs) [ 616.369919] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.369919] env[61857]: raise e [ 616.369919] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.369919] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 616.369919] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.369919] env[61857]: created_port_ids = self._update_ports_for_instance( [ 616.369919] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.369919] env[61857]: with excutils.save_and_reraise_exception(): [ 616.369919] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.369919] env[61857]: self.force_reraise() [ 616.369919] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.369919] env[61857]: raise self.value [ 616.369919] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.369919] env[61857]: updated_port = self._update_port( [ 616.369919] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.369919] env[61857]: _ensure_no_port_binding_failure(port) [ 616.369919] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.369919] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 616.370666] env[61857]: nova.exception.PortBindingFailed: Binding failed for port ea4ced50-70d2-45b3-b485-75acc9a5a73b, please check neutron logs for more information. [ 616.370666] env[61857]: Removing descriptor: 16 [ 616.370666] env[61857]: ERROR nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ea4ced50-70d2-45b3-b485-75acc9a5a73b, please check neutron logs for more information. [ 616.370666] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Traceback (most recent call last): [ 616.370666] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 616.370666] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] yield resources [ 616.370666] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.370666] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] self.driver.spawn(context, instance, image_meta, [ 616.370666] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 616.370666] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.370666] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.370666] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] vm_ref = self.build_virtual_machine(instance, [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] for vif in network_info: [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] return self._sync_wrapper(fn, *args, **kwargs) [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] self.wait() [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] self[:] = self._gt.wait() [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] return self._exit_event.wait() [ 616.370980] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] result = hub.switch() [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] return self.greenlet.switch() [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] result = function(*args, **kwargs) [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] return func(*args, **kwargs) [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] raise e [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] nwinfo = self.network_api.allocate_for_instance( [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.371393] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] created_port_ids = self._update_ports_for_instance( [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] with excutils.save_and_reraise_exception(): [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] self.force_reraise() [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] raise self.value [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] updated_port = self._update_port( [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] _ensure_no_port_binding_failure(port) [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.371713] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] raise exception.PortBindingFailed(port_id=port['id']) [ 616.372030] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] nova.exception.PortBindingFailed: Binding failed for port ea4ced50-70d2-45b3-b485-75acc9a5a73b, please check neutron logs for more information. [ 616.372030] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] [ 616.372030] env[61857]: INFO nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Terminating instance [ 616.998541] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Releasing lock "refresh_cache-0ae90dae-d095-418c-abda-f6bb76f89919" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.998801] env[61857]: DEBUG nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 616.998919] env[61857]: DEBUG nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.999090] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.002579] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "refresh_cache-b54af0ef-448e-453c-9056-32ad9141d9d3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.005895] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ac47e9-3a0a-4a6e-b8f2-b2ec81b1f1e0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.035485] env[61857]: DEBUG nova.network.neutron [req-d7c0e17b-e460-4e85-8e97-4d539cb12687 req-5a046a2a-7ebd-435e-8224-ae4de5d78b35 service nova] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.038553] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.040056] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a6bbac-c2fb-4732-be2c-2b66a1df320f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.049012] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d95e593-1d39-495f-9a76-ac8e9a34d0c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.064606] env[61857]: DEBUG nova.compute.provider_tree [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.245391] env[61857]: DEBUG nova.network.neutron [req-d7c0e17b-e460-4e85-8e97-4d539cb12687 req-5a046a2a-7ebd-435e-8224-ae4de5d78b35 service nova] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.544987] env[61857]: DEBUG nova.network.neutron [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.567319] env[61857]: DEBUG nova.scheduler.client.report [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.748449] env[61857]: DEBUG oslo_concurrency.lockutils [req-d7c0e17b-e460-4e85-8e97-4d539cb12687 req-5a046a2a-7ebd-435e-8224-ae4de5d78b35 service nova] Releasing lock "refresh_cache-b54af0ef-448e-453c-9056-32ad9141d9d3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.748916] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquired lock "refresh_cache-b54af0ef-448e-453c-9056-32ad9141d9d3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.749147] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.048068] env[61857]: INFO nova.compute.manager [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] [instance: 0ae90dae-d095-418c-abda-f6bb76f89919] Took 1.05 seconds to deallocate network for instance. [ 618.074031] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.728s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.074356] env[61857]: ERROR nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf, please check neutron logs for more information. [ 618.074356] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Traceback (most recent call last): [ 618.074356] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.074356] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] self.driver.spawn(context, instance, image_meta, [ 618.074356] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.074356] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.074356] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.074356] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] vm_ref = self.build_virtual_machine(instance, [ 618.074356] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.074356] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.074356] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] for vif in network_info: [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] return self._sync_wrapper(fn, *args, **kwargs) [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] self.wait() [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] self[:] = self._gt.wait() [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] return self._exit_event.wait() [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] result = hub.switch() [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.074655] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] return self.greenlet.switch() [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] result = function(*args, **kwargs) [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] return func(*args, **kwargs) [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] raise e [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] nwinfo = self.network_api.allocate_for_instance( [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] created_port_ids = self._update_ports_for_instance( [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] with excutils.save_and_reraise_exception(): [ 618.074950] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] self.force_reraise() [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] raise self.value [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] updated_port = self._update_port( [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] _ensure_no_port_binding_failure(port) [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] raise exception.PortBindingFailed(port_id=port['id']) [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] nova.exception.PortBindingFailed: Binding failed for port ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf, please check neutron logs for more information. [ 618.075254] env[61857]: ERROR nova.compute.manager [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] [ 618.075632] env[61857]: DEBUG nova.compute.utils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Binding failed for port ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 618.076936] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.160s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.080072] env[61857]: DEBUG nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Build of instance 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6 was re-scheduled: Binding failed for port ffcf558c-8ed6-45e5-9743-7ca5e2ef98cf, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 618.081754] env[61857]: DEBUG nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 618.082087] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Acquiring lock "refresh_cache-1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.082252] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Acquired lock "refresh_cache-1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.082506] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.267742] env[61857]: DEBUG nova.compute.manager [req-a1c4fc81-e6b2-4fc6-93f6-89d781947bb3 req-113626c3-6dac-41d7-8d41-db504b650e4a service nova] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Received event network-vif-deleted-ea4ced50-70d2-45b3-b485-75acc9a5a73b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 618.276870] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.404245] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.610196] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.683985] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.911133] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Releasing lock "refresh_cache-b54af0ef-448e-453c-9056-32ad9141d9d3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.911133] env[61857]: DEBUG nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 618.911133] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 618.911133] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2fd4d242-f9f8-424b-b426-5c0fb1031870 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.921310] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36540555-1822-47d2-8b7f-1c158ff81334 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.946827] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b54af0ef-448e-453c-9056-32ad9141d9d3 could not be found. [ 618.947689] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 618.947689] env[61857]: INFO nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 618.947689] env[61857]: DEBUG oslo.service.loopingcall [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.947890] env[61857]: DEBUG nova.compute.manager [-] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.948054] env[61857]: DEBUG nova.network.neutron [-] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 618.977728] env[61857]: DEBUG nova.network.neutron [-] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.057485] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1074578-fa33-4fe9-bb0d-91fb307a385c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.069652] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72a58176-fc43-4791-83fd-bb410e24384f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.104145] env[61857]: INFO nova.scheduler.client.report [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Deleted allocations for instance 0ae90dae-d095-418c-abda-f6bb76f89919 [ 619.109735] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea4d1f4-bcb7-420b-917b-111144753f6f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.120165] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e66f8ee-58f7-4b42-9625-69bae3572996 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.133961] env[61857]: DEBUG nova.compute.provider_tree [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.191885] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Releasing lock "refresh_cache-1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.192153] env[61857]: DEBUG nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 619.192323] env[61857]: DEBUG nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.192519] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.226992] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.479868] env[61857]: DEBUG nova.network.neutron [-] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.614915] env[61857]: DEBUG oslo_concurrency.lockutils [None req-450f3953-7e7c-4990-894f-0875e040abb9 tempest-ServerDiagnosticsTest-1845007325 tempest-ServerDiagnosticsTest-1845007325-project-member] Lock "0ae90dae-d095-418c-abda-f6bb76f89919" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.652s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.640263] env[61857]: DEBUG nova.scheduler.client.report [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.732615] env[61857]: DEBUG nova.network.neutron [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.982597] env[61857]: INFO nova.compute.manager [-] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Took 1.03 seconds to deallocate network for instance. [ 619.986133] env[61857]: DEBUG nova.compute.claims [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 619.986574] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.117577] env[61857]: DEBUG nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 620.148822] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.072s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.149509] env[61857]: ERROR nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e12d32c4-6bf0-422c-b988-cb04aa55820c, please check neutron logs for more information. [ 620.149509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Traceback (most recent call last): [ 620.149509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.149509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] self.driver.spawn(context, instance, image_meta, [ 620.149509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 620.149509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.149509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.149509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] vm_ref = self.build_virtual_machine(instance, [ 620.149509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.149509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.149509] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] for vif in network_info: [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] return self._sync_wrapper(fn, *args, **kwargs) [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] self.wait() [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] self[:] = self._gt.wait() [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] return self._exit_event.wait() [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] result = hub.switch() [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.149851] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] return self.greenlet.switch() [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] result = function(*args, **kwargs) [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] return func(*args, **kwargs) [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] raise e [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] nwinfo = self.network_api.allocate_for_instance( [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] created_port_ids = self._update_ports_for_instance( [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] with excutils.save_and_reraise_exception(): [ 620.150173] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] self.force_reraise() [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] raise self.value [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] updated_port = self._update_port( [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] _ensure_no_port_binding_failure(port) [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] raise exception.PortBindingFailed(port_id=port['id']) [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] nova.exception.PortBindingFailed: Binding failed for port e12d32c4-6bf0-422c-b988-cb04aa55820c, please check neutron logs for more information. [ 620.150476] env[61857]: ERROR nova.compute.manager [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] [ 620.150736] env[61857]: DEBUG nova.compute.utils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Binding failed for port e12d32c4-6bf0-422c-b988-cb04aa55820c, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 620.153411] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.636s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.154917] env[61857]: INFO nova.compute.claims [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 620.160179] env[61857]: DEBUG nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Build of instance f14ce7b5-158d-4831-9ba3-404b44752afa was re-scheduled: Binding failed for port e12d32c4-6bf0-422c-b988-cb04aa55820c, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 620.163020] env[61857]: DEBUG nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 620.163020] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "refresh_cache-f14ce7b5-158d-4831-9ba3-404b44752afa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.163020] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquired lock "refresh_cache-f14ce7b5-158d-4831-9ba3-404b44752afa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.163020] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 620.237205] env[61857]: INFO nova.compute.manager [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] [instance: 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6] Took 1.04 seconds to deallocate network for instance. [ 620.646510] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.711571] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.894489] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.275234] env[61857]: INFO nova.scheduler.client.report [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Deleted allocations for instance 1ca9a983-7ecf-44ac-b232-dd2ef19f77f6 [ 621.398114] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Releasing lock "refresh_cache-f14ce7b5-158d-4831-9ba3-404b44752afa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.398382] env[61857]: DEBUG nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 621.398565] env[61857]: DEBUG nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.398735] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 621.428897] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.676632] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9f1a1c-35c2-40b6-b2c2-e1839c8abc5a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.687894] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be349400-f3da-4c64-ab36-55511c8d4039 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.725265] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c65632-424e-4aeb-938f-10dd92ea2ad6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.735014] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fead630-4461-4438-afa9-70d5d966e8ef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.750666] env[61857]: DEBUG nova.compute.provider_tree [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.787443] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2cc1f79b-d2da-4f90-a741-78da56ea5af0 tempest-TenantUsagesTestJSON-392617592 tempest-TenantUsagesTestJSON-392617592-project-member] Lock "1ca9a983-7ecf-44ac-b232-dd2ef19f77f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.667s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.932596] env[61857]: DEBUG nova.network.neutron [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.254588] env[61857]: DEBUG nova.scheduler.client.report [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.291816] env[61857]: DEBUG nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 622.436542] env[61857]: INFO nova.compute.manager [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: f14ce7b5-158d-4831-9ba3-404b44752afa] Took 1.04 seconds to deallocate network for instance. [ 622.765021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.765021] env[61857]: DEBUG nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 622.766853] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.761s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.830757] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.272867] env[61857]: DEBUG nova.compute.utils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 623.278280] env[61857]: DEBUG nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 623.278280] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 623.346632] env[61857]: DEBUG nova.policy [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4fce1733cab54029a9a7328a94339fd1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06afaf24415b462bb36d93ed56fa2e5a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 623.480925] env[61857]: INFO nova.scheduler.client.report [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Deleted allocations for instance f14ce7b5-158d-4831-9ba3-404b44752afa [ 623.700415] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Successfully created port: 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.745530] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f141e1e-2e8e-469a-a50d-59ecd011a0ed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.753455] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2180d6-2cf7-45d2-a1e4-01c03caf13f9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.788028] env[61857]: DEBUG nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 623.791789] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a03833-155c-4c92-b229-23df8cbbfe03 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.801217] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a604326-ea67-4529-a5ac-ce38b6dca3e7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.821872] env[61857]: DEBUG nova.compute.provider_tree [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.991966] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72c4fc1a-ef12-4218-b375-b33c773ad777 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "f14ce7b5-158d-4831-9ba3-404b44752afa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.766s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.331446] env[61857]: DEBUG nova.scheduler.client.report [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.501021] env[61857]: DEBUG nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.805025] env[61857]: DEBUG nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 624.816945] env[61857]: DEBUG nova.compute.manager [req-ea65a963-7086-44b6-83a1-39a38edeed2b req-fec8fe18-5049-4e4d-899d-b9cd8419a386 service nova] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Received event network-changed-9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 624.816945] env[61857]: DEBUG nova.compute.manager [req-ea65a963-7086-44b6-83a1-39a38edeed2b req-fec8fe18-5049-4e4d-899d-b9cd8419a386 service nova] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Refreshing instance network info cache due to event network-changed-9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 624.817493] env[61857]: DEBUG oslo_concurrency.lockutils [req-ea65a963-7086-44b6-83a1-39a38edeed2b req-fec8fe18-5049-4e4d-899d-b9cd8419a386 service nova] Acquiring lock "refresh_cache-c1526323-a057-4a84-aa32-05cd802ebf9d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.818021] env[61857]: DEBUG oslo_concurrency.lockutils [req-ea65a963-7086-44b6-83a1-39a38edeed2b req-fec8fe18-5049-4e4d-899d-b9cd8419a386 service nova] Acquired lock "refresh_cache-c1526323-a057-4a84-aa32-05cd802ebf9d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.819701] env[61857]: DEBUG nova.network.neutron [req-ea65a963-7086-44b6-83a1-39a38edeed2b req-fec8fe18-5049-4e4d-899d-b9cd8419a386 service nova] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Refreshing network info cache for port 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 624.838034] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.069s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.838034] env[61857]: ERROR nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5551e212-0280-4973-8405-b66be8d415ae, please check neutron logs for more information. [ 624.838034] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Traceback (most recent call last): [ 624.838034] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.838034] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] self.driver.spawn(context, instance, image_meta, [ 624.838034] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 624.838034] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.838034] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.838034] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] vm_ref = self.build_virtual_machine(instance, [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] for vif in network_info: [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] return self._sync_wrapper(fn, *args, **kwargs) [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] self.wait() [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] self[:] = self._gt.wait() [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] return self._exit_event.wait() [ 624.838959] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] result = hub.switch() [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] return self.greenlet.switch() [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] result = function(*args, **kwargs) [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] return func(*args, **kwargs) [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] raise e [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] nwinfo = self.network_api.allocate_for_instance( [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.839637] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] created_port_ids = self._update_ports_for_instance( [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] with excutils.save_and_reraise_exception(): [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] self.force_reraise() [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] raise self.value [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] updated_port = self._update_port( [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] _ensure_no_port_binding_failure(port) [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.840252] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] raise exception.PortBindingFailed(port_id=port['id']) [ 624.840798] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] nova.exception.PortBindingFailed: Binding failed for port 5551e212-0280-4973-8405-b66be8d415ae, please check neutron logs for more information. [ 624.840798] env[61857]: ERROR nova.compute.manager [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] [ 624.840798] env[61857]: DEBUG nova.compute.utils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Binding failed for port 5551e212-0280-4973-8405-b66be8d415ae, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.847139] env[61857]: DEBUG nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Build of instance 5ad7fc93-72b3-4108-923d-f52d74589423 was re-scheduled: Binding failed for port 5551e212-0280-4973-8405-b66be8d415ae, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 624.847139] env[61857]: DEBUG nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 624.847139] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Acquiring lock "refresh_cache-5ad7fc93-72b3-4108-923d-f52d74589423" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.847139] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Acquired lock "refresh_cache-5ad7fc93-72b3-4108-923d-f52d74589423" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.847455] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 624.847455] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.030s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.847455] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.847455] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 624.847868] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.586s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.850587] env[61857]: INFO nova.compute.claims [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.862569] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422bc953-9fae-4c11-9214-1e4bf146447f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.881097] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 624.881476] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 624.881722] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.881990] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 624.882210] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.882440] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 624.882728] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 624.882942] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 624.883193] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 624.883426] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.883658] env[61857]: DEBUG nova.virt.hardware [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.885090] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87932d84-70c9-46fe-8bcc-deb0009c9ad5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.890603] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e2541d-8c3b-4ecc-836e-21fbbd8d02d4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.910010] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8dc094-c1bd-41e3-bacf-3cd409556924 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.914581] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c275975-a645-4916-b02a-2bc6be1bd7c4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.922457] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.939076] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d99a9c-a303-417d-8ede-88debdfa268a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.972132] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181643MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 624.972308] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.025056] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.071967] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.131339] env[61857]: ERROR nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa, please check neutron logs for more information. [ 625.131339] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 625.131339] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.131339] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 625.131339] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.131339] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 625.131339] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.131339] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 625.131339] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.131339] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 625.131339] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.131339] env[61857]: ERROR nova.compute.manager raise self.value [ 625.131339] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.131339] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 625.131339] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.131339] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 625.131881] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.131881] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 625.131881] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa, please check neutron logs for more information. [ 625.131881] env[61857]: ERROR nova.compute.manager [ 625.131881] env[61857]: Traceback (most recent call last): [ 625.131881] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 625.131881] env[61857]: listener.cb(fileno) [ 625.131881] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.131881] env[61857]: result = function(*args, **kwargs) [ 625.131881] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.131881] env[61857]: return func(*args, **kwargs) [ 625.131881] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.131881] env[61857]: raise e [ 625.131881] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.131881] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 625.131881] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.131881] env[61857]: created_port_ids = self._update_ports_for_instance( [ 625.131881] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.131881] env[61857]: with excutils.save_and_reraise_exception(): [ 625.131881] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.131881] env[61857]: self.force_reraise() [ 625.131881] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.131881] env[61857]: raise self.value [ 625.131881] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.131881] env[61857]: updated_port = self._update_port( [ 625.131881] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.131881] env[61857]: _ensure_no_port_binding_failure(port) [ 625.131881] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.131881] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 625.132573] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa, please check neutron logs for more information. [ 625.132573] env[61857]: Removing descriptor: 17 [ 625.132573] env[61857]: ERROR nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa, please check neutron logs for more information. [ 625.132573] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Traceback (most recent call last): [ 625.132573] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 625.132573] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] yield resources [ 625.132573] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.132573] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] self.driver.spawn(context, instance, image_meta, [ 625.132573] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 625.132573] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.132573] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.132573] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] vm_ref = self.build_virtual_machine(instance, [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] for vif in network_info: [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] return self._sync_wrapper(fn, *args, **kwargs) [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] self.wait() [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] self[:] = self._gt.wait() [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] return self._exit_event.wait() [ 625.132888] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] result = hub.switch() [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] return self.greenlet.switch() [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] result = function(*args, **kwargs) [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] return func(*args, **kwargs) [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] raise e [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] nwinfo = self.network_api.allocate_for_instance( [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.133266] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] created_port_ids = self._update_ports_for_instance( [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] with excutils.save_and_reraise_exception(): [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] self.force_reraise() [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] raise self.value [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] updated_port = self._update_port( [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] _ensure_no_port_binding_failure(port) [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.133613] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] raise exception.PortBindingFailed(port_id=port['id']) [ 625.133878] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] nova.exception.PortBindingFailed: Binding failed for port 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa, please check neutron logs for more information. [ 625.133878] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] [ 625.133878] env[61857]: INFO nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Terminating instance [ 625.133878] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Acquiring lock "refresh_cache-c1526323-a057-4a84-aa32-05cd802ebf9d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.352303] env[61857]: DEBUG nova.network.neutron [req-ea65a963-7086-44b6-83a1-39a38edeed2b req-fec8fe18-5049-4e4d-899d-b9cd8419a386 service nova] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.532378] env[61857]: DEBUG nova.network.neutron [req-ea65a963-7086-44b6-83a1-39a38edeed2b req-fec8fe18-5049-4e4d-899d-b9cd8419a386 service nova] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.574828] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Releasing lock "refresh_cache-5ad7fc93-72b3-4108-923d-f52d74589423" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.575081] env[61857]: DEBUG nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.575264] env[61857]: DEBUG nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.575428] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 625.601905] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.034458] env[61857]: DEBUG oslo_concurrency.lockutils [req-ea65a963-7086-44b6-83a1-39a38edeed2b req-fec8fe18-5049-4e4d-899d-b9cd8419a386 service nova] Releasing lock "refresh_cache-c1526323-a057-4a84-aa32-05cd802ebf9d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.035136] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Acquired lock "refresh_cache-c1526323-a057-4a84-aa32-05cd802ebf9d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.035468] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 626.110122] env[61857]: DEBUG nova.network.neutron [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.326030] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9fd3e24-a2a9-4324-909e-1bfafd1b36ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.336470] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2f0398-ba0c-439b-b71b-858a3cadb9e5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.378944] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60026f4-d255-426c-add6-3dcaad7bc81f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.388762] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affd7037-759f-44bf-8622-3d18ab23fb6a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.409198] env[61857]: DEBUG nova.compute.provider_tree [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.573682] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.613685] env[61857]: INFO nova.compute.manager [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] [instance: 5ad7fc93-72b3-4108-923d-f52d74589423] Took 1.04 seconds to deallocate network for instance. [ 626.751525] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.912478] env[61857]: DEBUG nova.scheduler.client.report [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.961871] env[61857]: DEBUG nova.compute.manager [req-21659e6e-01ff-432c-80c3-a7ee141007b9 req-2c0a97c9-c3bb-49ec-a58e-3f3c21618690 service nova] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Received event network-vif-deleted-9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 627.258221] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Releasing lock "refresh_cache-c1526323-a057-4a84-aa32-05cd802ebf9d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.258221] env[61857]: DEBUG nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 627.258221] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 627.258521] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-84dfa4a6-30ae-46fa-8c96-7bad472ab0d4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.272473] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d62fb2-13f3-48a1-9be4-23f639a4e59e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.295049] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c1526323-a057-4a84-aa32-05cd802ebf9d could not be found. [ 627.295401] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 627.295563] env[61857]: INFO nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 627.295711] env[61857]: DEBUG oslo.service.loopingcall [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 627.295929] env[61857]: DEBUG nova.compute.manager [-] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.296035] env[61857]: DEBUG nova.network.neutron [-] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 627.330252] env[61857]: DEBUG nova.network.neutron [-] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.427553] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.580s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.428106] env[61857]: DEBUG nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 627.431362] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.090s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.432957] env[61857]: INFO nova.compute.claims [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.659544] env[61857]: INFO nova.scheduler.client.report [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Deleted allocations for instance 5ad7fc93-72b3-4108-923d-f52d74589423 [ 627.833857] env[61857]: DEBUG nova.network.neutron [-] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.937802] env[61857]: DEBUG nova.compute.utils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.941300] env[61857]: DEBUG nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 627.941503] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 628.103031] env[61857]: DEBUG nova.policy [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3317265525e1499b87e847f2ad30ed60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '883ec0e28c9040dd89202cd07932ea7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.168377] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8b7da2e-da97-43fe-8c9e-b739943bd72f tempest-AttachInterfacesUnderV243Test-1377884608 tempest-AttachInterfacesUnderV243Test-1377884608-project-member] Lock "5ad7fc93-72b3-4108-923d-f52d74589423" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.749s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.338757] env[61857]: INFO nova.compute.manager [-] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Took 1.04 seconds to deallocate network for instance. [ 628.341943] env[61857]: DEBUG nova.compute.claims [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 628.342257] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.442287] env[61857]: DEBUG nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 628.671657] env[61857]: DEBUG nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.679966] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Successfully created port: a1f70ff1-faa9-40bd-9540-e9edfaabcdd9 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.931612] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d641ab6-bce4-42d8-9767-30496ddefe1b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.940505] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59907cc1-e063-4224-963d-a4d42411d213 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.990863] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a78a3c1-3727-41c5-a689-95e9b041278e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.001163] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ba10de-8c70-4f7e-87a6-77a98c0b28bf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.018108] env[61857]: DEBUG nova.compute.provider_tree [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.206384] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.455039] env[61857]: DEBUG nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 629.487172] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.487172] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.487172] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.487303] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.487303] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.487303] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.487303] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.487429] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.487543] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.487703] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.487872] env[61857]: DEBUG nova.virt.hardware [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.488790] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0536955-4b80-416b-9cd5-b74e09532fd1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.500672] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115a5cf0-550b-414b-a7cc-8f8d341ed901 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.521474] env[61857]: DEBUG nova.scheduler.client.report [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.027422] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.596s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.028184] env[61857]: DEBUG nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 630.036675] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.405s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.037747] env[61857]: INFO nova.compute.claims [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.199642] env[61857]: ERROR nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a1f70ff1-faa9-40bd-9540-e9edfaabcdd9, please check neutron logs for more information. [ 630.199642] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 630.199642] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.199642] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 630.199642] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.199642] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 630.199642] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.199642] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 630.199642] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.199642] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 630.199642] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.199642] env[61857]: ERROR nova.compute.manager raise self.value [ 630.199642] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.199642] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 630.199642] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.199642] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 630.200417] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.200417] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 630.200417] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a1f70ff1-faa9-40bd-9540-e9edfaabcdd9, please check neutron logs for more information. [ 630.200417] env[61857]: ERROR nova.compute.manager [ 630.200417] env[61857]: Traceback (most recent call last): [ 630.200417] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 630.200417] env[61857]: listener.cb(fileno) [ 630.200417] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.200417] env[61857]: result = function(*args, **kwargs) [ 630.200417] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.200417] env[61857]: return func(*args, **kwargs) [ 630.200417] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.200417] env[61857]: raise e [ 630.200417] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.200417] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 630.200417] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.200417] env[61857]: created_port_ids = self._update_ports_for_instance( [ 630.200417] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.200417] env[61857]: with excutils.save_and_reraise_exception(): [ 630.200417] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.200417] env[61857]: self.force_reraise() [ 630.200417] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.200417] env[61857]: raise self.value [ 630.200417] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.200417] env[61857]: updated_port = self._update_port( [ 630.200417] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.200417] env[61857]: _ensure_no_port_binding_failure(port) [ 630.200417] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.200417] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 630.201645] env[61857]: nova.exception.PortBindingFailed: Binding failed for port a1f70ff1-faa9-40bd-9540-e9edfaabcdd9, please check neutron logs for more information. [ 630.201645] env[61857]: Removing descriptor: 17 [ 630.201645] env[61857]: ERROR nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a1f70ff1-faa9-40bd-9540-e9edfaabcdd9, please check neutron logs for more information. [ 630.201645] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Traceback (most recent call last): [ 630.201645] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 630.201645] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] yield resources [ 630.201645] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.201645] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] self.driver.spawn(context, instance, image_meta, [ 630.201645] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 630.201645] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.201645] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.201645] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] vm_ref = self.build_virtual_machine(instance, [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] for vif in network_info: [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] return self._sync_wrapper(fn, *args, **kwargs) [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] self.wait() [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] self[:] = self._gt.wait() [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] return self._exit_event.wait() [ 630.202075] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] result = hub.switch() [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] return self.greenlet.switch() [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] result = function(*args, **kwargs) [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] return func(*args, **kwargs) [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] raise e [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] nwinfo = self.network_api.allocate_for_instance( [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.202625] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] created_port_ids = self._update_ports_for_instance( [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] with excutils.save_and_reraise_exception(): [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] self.force_reraise() [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] raise self.value [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] updated_port = self._update_port( [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] _ensure_no_port_binding_failure(port) [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.203072] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] raise exception.PortBindingFailed(port_id=port['id']) [ 630.203557] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] nova.exception.PortBindingFailed: Binding failed for port a1f70ff1-faa9-40bd-9540-e9edfaabcdd9, please check neutron logs for more information. [ 630.203557] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] [ 630.203557] env[61857]: INFO nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Terminating instance [ 630.203557] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquiring lock "refresh_cache-dafceac5-1eea-4103-8a48-ca6c5136390e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.203557] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquired lock "refresh_cache-dafceac5-1eea-4103-8a48-ca6c5136390e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.203557] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.236142] env[61857]: DEBUG nova.compute.manager [req-7c951eb6-7d56-4c5b-94cf-78cf3cca011b req-26cbe1ca-da7c-4f30-b9b4-6da61a2b5d5e service nova] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Received event network-changed-a1f70ff1-faa9-40bd-9540-e9edfaabcdd9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 630.236142] env[61857]: DEBUG nova.compute.manager [req-7c951eb6-7d56-4c5b-94cf-78cf3cca011b req-26cbe1ca-da7c-4f30-b9b4-6da61a2b5d5e service nova] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Refreshing instance network info cache due to event network-changed-a1f70ff1-faa9-40bd-9540-e9edfaabcdd9. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 630.236142] env[61857]: DEBUG oslo_concurrency.lockutils [req-7c951eb6-7d56-4c5b-94cf-78cf3cca011b req-26cbe1ca-da7c-4f30-b9b4-6da61a2b5d5e service nova] Acquiring lock "refresh_cache-dafceac5-1eea-4103-8a48-ca6c5136390e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.544931] env[61857]: DEBUG nova.compute.utils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 630.554018] env[61857]: DEBUG nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 630.554018] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 630.667013] env[61857]: DEBUG nova.policy [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df3958a0540d4b999d4e267daa8e392e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f026beafe7464c789744979bb8810adb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 630.729927] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.730212] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.739883] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.852241] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.054297] env[61857]: DEBUG nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 631.319957] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Successfully created port: c612b995-5fce-4231-a737-bdc5b3b06635 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 631.355264] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Releasing lock "refresh_cache-dafceac5-1eea-4103-8a48-ca6c5136390e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.355675] env[61857]: DEBUG nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.355869] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 631.358316] env[61857]: DEBUG oslo_concurrency.lockutils [req-7c951eb6-7d56-4c5b-94cf-78cf3cca011b req-26cbe1ca-da7c-4f30-b9b4-6da61a2b5d5e service nova] Acquired lock "refresh_cache-dafceac5-1eea-4103-8a48-ca6c5136390e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.358515] env[61857]: DEBUG nova.network.neutron [req-7c951eb6-7d56-4c5b-94cf-78cf3cca011b req-26cbe1ca-da7c-4f30-b9b4-6da61a2b5d5e service nova] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Refreshing network info cache for port a1f70ff1-faa9-40bd-9540-e9edfaabcdd9 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 631.359593] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a0002d3-0e43-487a-98dc-91b8a7997a87 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.376607] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf72de2-5fc2-4025-9c9a-2695146aed6f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.405257] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dafceac5-1eea-4103-8a48-ca6c5136390e could not be found. [ 631.405386] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 631.405554] env[61857]: INFO nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 631.405844] env[61857]: DEBUG oslo.service.loopingcall [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.406018] env[61857]: DEBUG nova.compute.manager [-] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.406132] env[61857]: DEBUG nova.network.neutron [-] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.436705] env[61857]: DEBUG nova.network.neutron [-] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.458922] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96800f7-3929-4b7a-ba23-84293f09adfb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.468530] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c96884-3819-4cdf-929d-49c6f01fee57 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.503945] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485c17db-4311-44ad-ad79-582ce64ced7a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.511997] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11afe63-9fb2-4d4b-8ca4-73056b3c3db7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.530234] env[61857]: DEBUG nova.compute.provider_tree [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.886174] env[61857]: DEBUG nova.network.neutron [req-7c951eb6-7d56-4c5b-94cf-78cf3cca011b req-26cbe1ca-da7c-4f30-b9b4-6da61a2b5d5e service nova] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.941252] env[61857]: DEBUG nova.network.neutron [-] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.987139] env[61857]: DEBUG nova.network.neutron [req-7c951eb6-7d56-4c5b-94cf-78cf3cca011b req-26cbe1ca-da7c-4f30-b9b4-6da61a2b5d5e service nova] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.036613] env[61857]: DEBUG nova.scheduler.client.report [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.071182] env[61857]: DEBUG nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 632.100895] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.101166] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.101372] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.101506] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.101652] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.101800] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.102060] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.102171] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.102339] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.102500] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.102674] env[61857]: DEBUG nova.virt.hardware [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.103546] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955263c7-6b2a-4043-80a8-78e1008ae0fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.113258] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f8a689-c5af-4348-8d5d-7d1e7ce41df2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.303503] env[61857]: DEBUG nova.compute.manager [req-5727590a-a5b7-435d-9d27-2cc37765057d req-8a505896-2e6d-4aaf-8f80-7af48ebf6db5 service nova] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Received event network-vif-deleted-a1f70ff1-faa9-40bd-9540-e9edfaabcdd9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 632.444341] env[61857]: INFO nova.compute.manager [-] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Took 1.04 seconds to deallocate network for instance. [ 632.447069] env[61857]: DEBUG nova.compute.claims [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.447354] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.491171] env[61857]: DEBUG oslo_concurrency.lockutils [req-7c951eb6-7d56-4c5b-94cf-78cf3cca011b req-26cbe1ca-da7c-4f30-b9b4-6da61a2b5d5e service nova] Releasing lock "refresh_cache-dafceac5-1eea-4103-8a48-ca6c5136390e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.541828] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.542708] env[61857]: DEBUG nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 632.549111] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.026s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.549111] env[61857]: INFO nova.compute.claims [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.729015] env[61857]: ERROR nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c612b995-5fce-4231-a737-bdc5b3b06635, please check neutron logs for more information. [ 632.729015] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.729015] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.729015] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.729015] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.729015] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.729015] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.729015] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.729015] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.729015] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 632.729015] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.729015] env[61857]: ERROR nova.compute.manager raise self.value [ 632.729015] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.729015] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.729015] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.729015] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.732380] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.732380] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.732380] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c612b995-5fce-4231-a737-bdc5b3b06635, please check neutron logs for more information. [ 632.732380] env[61857]: ERROR nova.compute.manager [ 632.732380] env[61857]: Traceback (most recent call last): [ 632.732380] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.732380] env[61857]: listener.cb(fileno) [ 632.732380] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.732380] env[61857]: result = function(*args, **kwargs) [ 632.732380] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.732380] env[61857]: return func(*args, **kwargs) [ 632.732380] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.732380] env[61857]: raise e [ 632.732380] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.732380] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 632.732380] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.732380] env[61857]: created_port_ids = self._update_ports_for_instance( [ 632.732380] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.732380] env[61857]: with excutils.save_and_reraise_exception(): [ 632.732380] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.732380] env[61857]: self.force_reraise() [ 632.732380] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.732380] env[61857]: raise self.value [ 632.732380] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.732380] env[61857]: updated_port = self._update_port( [ 632.732380] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.732380] env[61857]: _ensure_no_port_binding_failure(port) [ 632.732380] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.732380] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.734021] env[61857]: nova.exception.PortBindingFailed: Binding failed for port c612b995-5fce-4231-a737-bdc5b3b06635, please check neutron logs for more information. [ 632.734021] env[61857]: Removing descriptor: 17 [ 632.734021] env[61857]: ERROR nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c612b995-5fce-4231-a737-bdc5b3b06635, please check neutron logs for more information. [ 632.734021] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] Traceback (most recent call last): [ 632.734021] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 632.734021] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] yield resources [ 632.734021] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.734021] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] self.driver.spawn(context, instance, image_meta, [ 632.734021] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.734021] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.734021] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.734021] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] vm_ref = self.build_virtual_machine(instance, [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] for vif in network_info: [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] return self._sync_wrapper(fn, *args, **kwargs) [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] self.wait() [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] self[:] = self._gt.wait() [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] return self._exit_event.wait() [ 632.734299] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] result = hub.switch() [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] return self.greenlet.switch() [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] result = function(*args, **kwargs) [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] return func(*args, **kwargs) [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] raise e [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] nwinfo = self.network_api.allocate_for_instance( [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.734619] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] created_port_ids = self._update_ports_for_instance( [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] with excutils.save_and_reraise_exception(): [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] self.force_reraise() [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] raise self.value [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] updated_port = self._update_port( [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] _ensure_no_port_binding_failure(port) [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.734943] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] raise exception.PortBindingFailed(port_id=port['id']) [ 632.735314] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] nova.exception.PortBindingFailed: Binding failed for port c612b995-5fce-4231-a737-bdc5b3b06635, please check neutron logs for more information. [ 632.735314] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] [ 632.735314] env[61857]: INFO nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Terminating instance [ 632.735314] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "refresh_cache-406f6726-4009-49bd-a37b-91707ece1932" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.735314] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquired lock "refresh_cache-406f6726-4009-49bd-a37b-91707ece1932" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.735314] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.047966] env[61857]: DEBUG nova.compute.utils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.049412] env[61857]: DEBUG nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 633.051097] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 633.118921] env[61857]: DEBUG nova.policy [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db4e4c99c7b64a45a9f483ac583cf61b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f32c316491d34a78a69a8c3f68d3dc8a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 633.257603] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.416739] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "d49008e6-7789-4785-8eea-9b3fa36c4ccb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.417119] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "d49008e6-7789-4785-8eea-9b3fa36c4ccb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.425513] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.559201] env[61857]: DEBUG nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 633.569300] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Successfully created port: ae0205a7-5adf-4aae-a937-72a665c82f24 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.931946] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Releasing lock "refresh_cache-406f6726-4009-49bd-a37b-91707ece1932" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.932421] env[61857]: DEBUG nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.932806] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 633.934023] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "3cacef7b-dd24-4d33-9500-bbff03e342bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.934023] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "3cacef7b-dd24-4d33-9500-bbff03e342bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.934814] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1937b97c-828d-4dc1-b38c-c0cf134c67b4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.947638] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be50157-a996-4d55-bc3c-d8f917f2d8b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.972901] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 406f6726-4009-49bd-a37b-91707ece1932 could not be found. [ 633.973298] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 633.973387] env[61857]: INFO nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Took 0.04 seconds to destroy the instance on the hypervisor. [ 633.974541] env[61857]: DEBUG oslo.service.loopingcall [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.974541] env[61857]: DEBUG nova.compute.manager [-] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.974541] env[61857]: DEBUG nova.network.neutron [-] [instance: 406f6726-4009-49bd-a37b-91707ece1932] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.019032] env[61857]: DEBUG nova.network.neutron [-] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.082602] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc32307-7a1f-4c92-bbd4-49d4c42cfd78 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.088482] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc6eb91-255a-4203-89f6-bb0fb58902b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.124018] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a46d4c-05a3-46b0-aeb5-ef07a65a2b2a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.133192] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5accc0-dafc-4fa7-9319-0f99792368a2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.146176] env[61857]: DEBUG nova.compute.provider_tree [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.328474] env[61857]: DEBUG nova.compute.manager [req-04c982fe-a94a-4975-a50f-5d003be0f406 req-16dd01e8-e742-4a21-8cc4-2bbf7ab18e37 service nova] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Received event network-changed-c612b995-5fce-4231-a737-bdc5b3b06635 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 634.328666] env[61857]: DEBUG nova.compute.manager [req-04c982fe-a94a-4975-a50f-5d003be0f406 req-16dd01e8-e742-4a21-8cc4-2bbf7ab18e37 service nova] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Refreshing instance network info cache due to event network-changed-c612b995-5fce-4231-a737-bdc5b3b06635. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 634.328875] env[61857]: DEBUG oslo_concurrency.lockutils [req-04c982fe-a94a-4975-a50f-5d003be0f406 req-16dd01e8-e742-4a21-8cc4-2bbf7ab18e37 service nova] Acquiring lock "refresh_cache-406f6726-4009-49bd-a37b-91707ece1932" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.329027] env[61857]: DEBUG oslo_concurrency.lockutils [req-04c982fe-a94a-4975-a50f-5d003be0f406 req-16dd01e8-e742-4a21-8cc4-2bbf7ab18e37 service nova] Acquired lock "refresh_cache-406f6726-4009-49bd-a37b-91707ece1932" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.329188] env[61857]: DEBUG nova.network.neutron [req-04c982fe-a94a-4975-a50f-5d003be0f406 req-16dd01e8-e742-4a21-8cc4-2bbf7ab18e37 service nova] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Refreshing network info cache for port c612b995-5fce-4231-a737-bdc5b3b06635 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 634.521896] env[61857]: DEBUG nova.network.neutron [-] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.575282] env[61857]: DEBUG nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 634.607105] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 634.607105] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 634.607105] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.607285] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 634.607285] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.607285] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 634.607285] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 634.607285] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 634.607858] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 634.608215] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 634.608568] env[61857]: DEBUG nova.virt.hardware [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.609559] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b2c9fb-6898-412e-bbcf-0bd1d3f32bcc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.620015] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6483fd41-c13b-4782-87f7-8e3f8a5b0517 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.648764] env[61857]: DEBUG nova.scheduler.client.report [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.855053] env[61857]: DEBUG nova.network.neutron [req-04c982fe-a94a-4975-a50f-5d003be0f406 req-16dd01e8-e742-4a21-8cc4-2bbf7ab18e37 service nova] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.952309] env[61857]: DEBUG nova.network.neutron [req-04c982fe-a94a-4975-a50f-5d003be0f406 req-16dd01e8-e742-4a21-8cc4-2bbf7ab18e37 service nova] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.027936] env[61857]: INFO nova.compute.manager [-] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Took 1.05 seconds to deallocate network for instance. [ 635.038579] env[61857]: DEBUG nova.compute.claims [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.038988] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.154033] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.608s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.154589] env[61857]: DEBUG nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 635.161794] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.175s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.458695] env[61857]: DEBUG oslo_concurrency.lockutils [req-04c982fe-a94a-4975-a50f-5d003be0f406 req-16dd01e8-e742-4a21-8cc4-2bbf7ab18e37 service nova] Releasing lock "refresh_cache-406f6726-4009-49bd-a37b-91707ece1932" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.459076] env[61857]: DEBUG nova.compute.manager [req-04c982fe-a94a-4975-a50f-5d003be0f406 req-16dd01e8-e742-4a21-8cc4-2bbf7ab18e37 service nova] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Received event network-vif-deleted-c612b995-5fce-4231-a737-bdc5b3b06635 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 635.532212] env[61857]: ERROR nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ae0205a7-5adf-4aae-a937-72a665c82f24, please check neutron logs for more information. [ 635.532212] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.532212] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.532212] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.532212] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.532212] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.532212] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.532212] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.532212] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.532212] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 635.532212] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.532212] env[61857]: ERROR nova.compute.manager raise self.value [ 635.532212] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.532212] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.532212] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.532212] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.532594] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.532594] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.532594] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ae0205a7-5adf-4aae-a937-72a665c82f24, please check neutron logs for more information. [ 635.532594] env[61857]: ERROR nova.compute.manager [ 635.532594] env[61857]: Traceback (most recent call last): [ 635.532594] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.532594] env[61857]: listener.cb(fileno) [ 635.532594] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.532594] env[61857]: result = function(*args, **kwargs) [ 635.532594] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.532594] env[61857]: return func(*args, **kwargs) [ 635.532594] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.532594] env[61857]: raise e [ 635.532594] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.532594] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 635.532594] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.532594] env[61857]: created_port_ids = self._update_ports_for_instance( [ 635.532594] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.532594] env[61857]: with excutils.save_and_reraise_exception(): [ 635.532594] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.532594] env[61857]: self.force_reraise() [ 635.532594] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.532594] env[61857]: raise self.value [ 635.532594] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.532594] env[61857]: updated_port = self._update_port( [ 635.532594] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.532594] env[61857]: _ensure_no_port_binding_failure(port) [ 635.532594] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.532594] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.533260] env[61857]: nova.exception.PortBindingFailed: Binding failed for port ae0205a7-5adf-4aae-a937-72a665c82f24, please check neutron logs for more information. [ 635.533260] env[61857]: Removing descriptor: 17 [ 635.533260] env[61857]: ERROR nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ae0205a7-5adf-4aae-a937-72a665c82f24, please check neutron logs for more information. [ 635.533260] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Traceback (most recent call last): [ 635.533260] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.533260] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] yield resources [ 635.533260] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.533260] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] self.driver.spawn(context, instance, image_meta, [ 635.533260] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.533260] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.533260] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.533260] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] vm_ref = self.build_virtual_machine(instance, [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] for vif in network_info: [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] return self._sync_wrapper(fn, *args, **kwargs) [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] self.wait() [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] self[:] = self._gt.wait() [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] return self._exit_event.wait() [ 635.533651] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] result = hub.switch() [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] return self.greenlet.switch() [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] result = function(*args, **kwargs) [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] return func(*args, **kwargs) [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] raise e [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] nwinfo = self.network_api.allocate_for_instance( [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.533966] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] created_port_ids = self._update_ports_for_instance( [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] with excutils.save_and_reraise_exception(): [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] self.force_reraise() [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] raise self.value [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] updated_port = self._update_port( [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] _ensure_no_port_binding_failure(port) [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.534282] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] raise exception.PortBindingFailed(port_id=port['id']) [ 635.534559] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] nova.exception.PortBindingFailed: Binding failed for port ae0205a7-5adf-4aae-a937-72a665c82f24, please check neutron logs for more information. [ 635.534559] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] [ 635.534559] env[61857]: INFO nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Terminating instance [ 635.537335] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Acquiring lock "refresh_cache-3375702e-36fd-43d4-95da-e3aafb96a32f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.538154] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Acquired lock "refresh_cache-3375702e-36fd-43d4-95da-e3aafb96a32f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.538326] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.668275] env[61857]: DEBUG nova.compute.utils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 635.669723] env[61857]: DEBUG nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 635.669902] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 635.717873] env[61857]: DEBUG nova.policy [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f4de145d44e4c55b2ed57ce90bcd88e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9261dd9ab392430eaea9d5a9b00a2705', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.029780] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "e286e6d0-25db-4d86-b9a2-3538bc3f624a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.030125] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "e286e6d0-25db-4d86-b9a2-3538bc3f624a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.085953] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.144625] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Successfully created port: 20789189-3d73-49fc-a0ed-04d7cdd3cb78 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.175802] env[61857]: DEBUG nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 636.206614] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bed66f9-53f1-4dc2-a69a-794ccfbc8e41 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.217660] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d311eb51-69bb-43ab-8c27-c9c3d6fe02db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.255747] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9911a013-1a00-4390-891a-4197c90afcb6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.266089] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847c9f7f-b46f-44ae-92f2-17c4a65d8b73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.282080] env[61857]: DEBUG nova.compute.provider_tree [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.364941] env[61857]: DEBUG nova.compute.manager [req-1109e521-9300-44d2-a640-837c2a62ffd1 req-a4eab767-ea15-46f3-ba6f-7f94b33e6099 service nova] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Received event network-changed-ae0205a7-5adf-4aae-a937-72a665c82f24 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 636.365163] env[61857]: DEBUG nova.compute.manager [req-1109e521-9300-44d2-a640-837c2a62ffd1 req-a4eab767-ea15-46f3-ba6f-7f94b33e6099 service nova] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Refreshing instance network info cache due to event network-changed-ae0205a7-5adf-4aae-a937-72a665c82f24. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 636.365779] env[61857]: DEBUG oslo_concurrency.lockutils [req-1109e521-9300-44d2-a640-837c2a62ffd1 req-a4eab767-ea15-46f3-ba6f-7f94b33e6099 service nova] Acquiring lock "refresh_cache-3375702e-36fd-43d4-95da-e3aafb96a32f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.453572] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.784873] env[61857]: DEBUG nova.scheduler.client.report [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.956765] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Releasing lock "refresh_cache-3375702e-36fd-43d4-95da-e3aafb96a32f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.957249] env[61857]: DEBUG nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.957458] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 636.957794] env[61857]: DEBUG oslo_concurrency.lockutils [req-1109e521-9300-44d2-a640-837c2a62ffd1 req-a4eab767-ea15-46f3-ba6f-7f94b33e6099 service nova] Acquired lock "refresh_cache-3375702e-36fd-43d4-95da-e3aafb96a32f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.957972] env[61857]: DEBUG nova.network.neutron [req-1109e521-9300-44d2-a640-837c2a62ffd1 req-a4eab767-ea15-46f3-ba6f-7f94b33e6099 service nova] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Refreshing network info cache for port ae0205a7-5adf-4aae-a937-72a665c82f24 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 636.959132] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c50905be-3ac5-4b0f-8095-6fae6aecbbf1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.974333] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59841fb3-ea62-4301-82e7-98ce2eea076f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.003339] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3375702e-36fd-43d4-95da-e3aafb96a32f could not be found. [ 637.003597] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 637.003780] env[61857]: INFO nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 637.007451] env[61857]: DEBUG oslo.service.loopingcall [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.007451] env[61857]: DEBUG nova.compute.manager [-] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.007451] env[61857]: DEBUG nova.network.neutron [-] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 637.022504] env[61857]: DEBUG nova.network.neutron [-] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.185772] env[61857]: DEBUG nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 637.220499] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.220499] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.220499] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.220707] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.220707] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.220707] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.220909] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.221085] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.221251] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.221407] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.221577] env[61857]: DEBUG nova.virt.hardware [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.222467] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cafc5bd-d9c1-44b0-adda-90812a1d7396 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.233339] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f48745-466a-4718-9cfe-440338b4f8c0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.296033] env[61857]: ERROR nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 20789189-3d73-49fc-a0ed-04d7cdd3cb78, please check neutron logs for more information. [ 637.296033] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.296033] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.296033] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.296033] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.296033] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.296033] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.296033] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.296033] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.296033] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 637.296033] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.296033] env[61857]: ERROR nova.compute.manager raise self.value [ 637.296033] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.296033] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.296033] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.296033] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.296473] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.296473] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.296473] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 20789189-3d73-49fc-a0ed-04d7cdd3cb78, please check neutron logs for more information. [ 637.296473] env[61857]: ERROR nova.compute.manager [ 637.296473] env[61857]: Traceback (most recent call last): [ 637.296473] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.296473] env[61857]: listener.cb(fileno) [ 637.296473] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.296473] env[61857]: result = function(*args, **kwargs) [ 637.296473] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.296473] env[61857]: return func(*args, **kwargs) [ 637.296473] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.296473] env[61857]: raise e [ 637.296473] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.296473] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 637.296473] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.296473] env[61857]: created_port_ids = self._update_ports_for_instance( [ 637.296473] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.296473] env[61857]: with excutils.save_and_reraise_exception(): [ 637.296473] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.296473] env[61857]: self.force_reraise() [ 637.296473] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.296473] env[61857]: raise self.value [ 637.296473] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.296473] env[61857]: updated_port = self._update_port( [ 637.296473] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.296473] env[61857]: _ensure_no_port_binding_failure(port) [ 637.296473] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.296473] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.297166] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 20789189-3d73-49fc-a0ed-04d7cdd3cb78, please check neutron logs for more information. [ 637.297166] env[61857]: Removing descriptor: 17 [ 637.297166] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.135s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.297166] env[61857]: ERROR nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ea4ced50-70d2-45b3-b485-75acc9a5a73b, please check neutron logs for more information. [ 637.297166] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Traceback (most recent call last): [ 637.297166] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.297166] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] self.driver.spawn(context, instance, image_meta, [ 637.297166] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 637.297166] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.297166] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] vm_ref = self.build_virtual_machine(instance, [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] for vif in network_info: [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] return self._sync_wrapper(fn, *args, **kwargs) [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] self.wait() [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] self[:] = self._gt.wait() [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.297498] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] return self._exit_event.wait() [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] result = hub.switch() [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] return self.greenlet.switch() [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] result = function(*args, **kwargs) [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] return func(*args, **kwargs) [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] raise e [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] nwinfo = self.network_api.allocate_for_instance( [ 637.297795] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] created_port_ids = self._update_ports_for_instance( [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] with excutils.save_and_reraise_exception(): [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] self.force_reraise() [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] raise self.value [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] updated_port = self._update_port( [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] _ensure_no_port_binding_failure(port) [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.298299] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] raise exception.PortBindingFailed(port_id=port['id']) [ 637.298607] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] nova.exception.PortBindingFailed: Binding failed for port ea4ced50-70d2-45b3-b485-75acc9a5a73b, please check neutron logs for more information. [ 637.298607] env[61857]: ERROR nova.compute.manager [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] [ 637.298607] env[61857]: DEBUG nova.compute.utils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Binding failed for port ea4ced50-70d2-45b3-b485-75acc9a5a73b, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 637.300621] env[61857]: ERROR nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 20789189-3d73-49fc-a0ed-04d7cdd3cb78, please check neutron logs for more information. [ 637.300621] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Traceback (most recent call last): [ 637.300621] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 637.300621] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] yield resources [ 637.300621] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.300621] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] self.driver.spawn(context, instance, image_meta, [ 637.300621] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 637.300621] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.300621] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.300621] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] vm_ref = self.build_virtual_machine(instance, [ 637.300621] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] for vif in network_info: [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] return self._sync_wrapper(fn, *args, **kwargs) [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] self.wait() [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] self[:] = self._gt.wait() [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] return self._exit_event.wait() [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.300966] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] result = hub.switch() [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] return self.greenlet.switch() [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] result = function(*args, **kwargs) [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] return func(*args, **kwargs) [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] raise e [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] nwinfo = self.network_api.allocate_for_instance( [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] created_port_ids = self._update_ports_for_instance( [ 637.301424] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] with excutils.save_and_reraise_exception(): [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] self.force_reraise() [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] raise self.value [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] updated_port = self._update_port( [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] _ensure_no_port_binding_failure(port) [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] raise exception.PortBindingFailed(port_id=port['id']) [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] nova.exception.PortBindingFailed: Binding failed for port 20789189-3d73-49fc-a0ed-04d7cdd3cb78, please check neutron logs for more information. [ 637.302306] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] [ 637.302863] env[61857]: INFO nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Terminating instance [ 637.302863] env[61857]: DEBUG nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Build of instance b54af0ef-448e-453c-9056-32ad9141d9d3 was re-scheduled: Binding failed for port ea4ced50-70d2-45b3-b485-75acc9a5a73b, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 637.302863] env[61857]: DEBUG nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 637.302863] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "refresh_cache-b54af0ef-448e-453c-9056-32ad9141d9d3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.302863] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquired lock "refresh_cache-b54af0ef-448e-453c-9056-32ad9141d9d3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.303146] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 637.303832] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.657s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.306366] env[61857]: INFO nova.compute.claims [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.320283] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Acquiring lock "refresh_cache-1492bf97-9e03-40eb-ae6c-041f2998124e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.322905] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Acquired lock "refresh_cache-1492bf97-9e03-40eb-ae6c-041f2998124e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.322905] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 637.487146] env[61857]: DEBUG nova.network.neutron [req-1109e521-9300-44d2-a640-837c2a62ffd1 req-a4eab767-ea15-46f3-ba6f-7f94b33e6099 service nova] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.525427] env[61857]: DEBUG nova.network.neutron [-] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.580676] env[61857]: DEBUG nova.network.neutron [req-1109e521-9300-44d2-a640-837c2a62ffd1 req-a4eab767-ea15-46f3-ba6f-7f94b33e6099 service nova] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.608392] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquiring lock "7a800abe-ea0a-4077-a5da-dd60eab917e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.608961] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lock "7a800abe-ea0a-4077-a5da-dd60eab917e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.833357] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.845044] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.929250] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.945368] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.030304] env[61857]: INFO nova.compute.manager [-] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Took 1.03 seconds to deallocate network for instance. [ 638.032707] env[61857]: DEBUG nova.compute.claims [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.032887] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.084698] env[61857]: DEBUG oslo_concurrency.lockutils [req-1109e521-9300-44d2-a640-837c2a62ffd1 req-a4eab767-ea15-46f3-ba6f-7f94b33e6099 service nova] Releasing lock "refresh_cache-3375702e-36fd-43d4-95da-e3aafb96a32f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.085009] env[61857]: DEBUG nova.compute.manager [req-1109e521-9300-44d2-a640-837c2a62ffd1 req-a4eab767-ea15-46f3-ba6f-7f94b33e6099 service nova] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Received event network-vif-deleted-ae0205a7-5adf-4aae-a937-72a665c82f24 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 638.415817] env[61857]: DEBUG nova.compute.manager [req-b6c7be59-ff33-4161-81af-b8ba7661e5d0 req-0c6ab696-6403-428f-89e0-043db197eec5 service nova] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Received event network-changed-20789189-3d73-49fc-a0ed-04d7cdd3cb78 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 638.416115] env[61857]: DEBUG nova.compute.manager [req-b6c7be59-ff33-4161-81af-b8ba7661e5d0 req-0c6ab696-6403-428f-89e0-043db197eec5 service nova] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Refreshing instance network info cache due to event network-changed-20789189-3d73-49fc-a0ed-04d7cdd3cb78. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 638.416304] env[61857]: DEBUG oslo_concurrency.lockutils [req-b6c7be59-ff33-4161-81af-b8ba7661e5d0 req-0c6ab696-6403-428f-89e0-043db197eec5 service nova] Acquiring lock "refresh_cache-1492bf97-9e03-40eb-ae6c-041f2998124e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.433464] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Releasing lock "refresh_cache-b54af0ef-448e-453c-9056-32ad9141d9d3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.433693] env[61857]: DEBUG nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 638.433871] env[61857]: DEBUG nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.434445] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.447386] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Releasing lock "refresh_cache-1492bf97-9e03-40eb-ae6c-041f2998124e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.447764] env[61857]: DEBUG nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 638.447944] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 638.448464] env[61857]: DEBUG oslo_concurrency.lockutils [req-b6c7be59-ff33-4161-81af-b8ba7661e5d0 req-0c6ab696-6403-428f-89e0-043db197eec5 service nova] Acquired lock "refresh_cache-1492bf97-9e03-40eb-ae6c-041f2998124e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.448631] env[61857]: DEBUG nova.network.neutron [req-b6c7be59-ff33-4161-81af-b8ba7661e5d0 req-0c6ab696-6403-428f-89e0-043db197eec5 service nova] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Refreshing network info cache for port 20789189-3d73-49fc-a0ed-04d7cdd3cb78 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 638.450458] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-88e13f3d-cea8-46f9-9d2e-35a96a097cd5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.453427] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.463644] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f559cc08-94e6-4365-9a75-11483204ea9a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.489347] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1492bf97-9e03-40eb-ae6c-041f2998124e could not be found. [ 638.489615] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 638.489809] env[61857]: INFO nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 638.490069] env[61857]: DEBUG oslo.service.loopingcall [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.493237] env[61857]: DEBUG nova.compute.manager [-] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.493337] env[61857]: DEBUG nova.network.neutron [-] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.517672] env[61857]: DEBUG nova.network.neutron [-] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.738795] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7982b313-b2dd-4fc7-be1b-919b970b0554 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.746773] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d609b7-f2d4-4fe8-bd06-a5393c7bb9d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.776977] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af7bd2a-fda8-46ec-932f-716b92380bcd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.784380] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9947fe76-ee48-466c-9e69-6a375112b675 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.797665] env[61857]: DEBUG nova.compute.provider_tree [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.957188] env[61857]: DEBUG nova.network.neutron [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.978751] env[61857]: DEBUG nova.network.neutron [req-b6c7be59-ff33-4161-81af-b8ba7661e5d0 req-0c6ab696-6403-428f-89e0-043db197eec5 service nova] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.022646] env[61857]: DEBUG nova.network.neutron [-] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.073639] env[61857]: DEBUG nova.network.neutron [req-b6c7be59-ff33-4161-81af-b8ba7661e5d0 req-0c6ab696-6403-428f-89e0-043db197eec5 service nova] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.300522] env[61857]: DEBUG nova.scheduler.client.report [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.461283] env[61857]: INFO nova.compute.manager [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: b54af0ef-448e-453c-9056-32ad9141d9d3] Took 1.03 seconds to deallocate network for instance. [ 639.523852] env[61857]: INFO nova.compute.manager [-] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Took 1.03 seconds to deallocate network for instance. [ 639.526257] env[61857]: DEBUG nova.compute.claims [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 639.526438] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.576796] env[61857]: DEBUG oslo_concurrency.lockutils [req-b6c7be59-ff33-4161-81af-b8ba7661e5d0 req-0c6ab696-6403-428f-89e0-043db197eec5 service nova] Releasing lock "refresh_cache-1492bf97-9e03-40eb-ae6c-041f2998124e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.576901] env[61857]: DEBUG nova.compute.manager [req-b6c7be59-ff33-4161-81af-b8ba7661e5d0 req-0c6ab696-6403-428f-89e0-043db197eec5 service nova] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Received event network-vif-deleted-20789189-3d73-49fc-a0ed-04d7cdd3cb78 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 639.805223] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.805755] env[61857]: DEBUG nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 639.808344] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.978s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.809720] env[61857]: INFO nova.compute.claims [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.314450] env[61857]: DEBUG nova.compute.utils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.317962] env[61857]: DEBUG nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 640.318656] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 640.366233] env[61857]: DEBUG nova.policy [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8722a9ebebd14eacac59c7812251de79', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd59a31d1f9740fb8f653b4406729a3a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.501582] env[61857]: INFO nova.scheduler.client.report [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Deleted allocations for instance b54af0ef-448e-453c-9056-32ad9141d9d3 [ 640.700472] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Successfully created port: c382853b-2a41-415a-918c-d07424264041 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 640.819115] env[61857]: DEBUG nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 641.018021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-579abe3d-a848-4f87-9dab-0c9f725f861d tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "b54af0ef-448e-453c-9056-32ad9141d9d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.789s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.241347] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fed41b8-7e58-4527-9f6d-b2db7da6fe5f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.249749] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84cf705e-d419-47b6-bc45-72f9ff106665 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.286102] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b8d651-25c6-47e7-922e-b826c8b7e685 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.294519] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432d8c33-0be0-407f-99f2-5d15fde3349b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.312718] env[61857]: DEBUG nova.compute.provider_tree [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.519628] env[61857]: DEBUG nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 641.608336] env[61857]: DEBUG nova.compute.manager [req-a7332305-b8b2-49ab-a209-b34d302b7e0c req-9deef5ce-fb87-4f72-83d3-33336d3884dc service nova] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Received event network-changed-c382853b-2a41-415a-918c-d07424264041 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 641.608520] env[61857]: DEBUG nova.compute.manager [req-a7332305-b8b2-49ab-a209-b34d302b7e0c req-9deef5ce-fb87-4f72-83d3-33336d3884dc service nova] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Refreshing instance network info cache due to event network-changed-c382853b-2a41-415a-918c-d07424264041. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 641.608732] env[61857]: DEBUG oslo_concurrency.lockutils [req-a7332305-b8b2-49ab-a209-b34d302b7e0c req-9deef5ce-fb87-4f72-83d3-33336d3884dc service nova] Acquiring lock "refresh_cache-ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.608878] env[61857]: DEBUG oslo_concurrency.lockutils [req-a7332305-b8b2-49ab-a209-b34d302b7e0c req-9deef5ce-fb87-4f72-83d3-33336d3884dc service nova] Acquired lock "refresh_cache-ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.610977] env[61857]: DEBUG nova.network.neutron [req-a7332305-b8b2-49ab-a209-b34d302b7e0c req-9deef5ce-fb87-4f72-83d3-33336d3884dc service nova] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Refreshing network info cache for port c382853b-2a41-415a-918c-d07424264041 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 641.760789] env[61857]: ERROR nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c382853b-2a41-415a-918c-d07424264041, please check neutron logs for more information. [ 641.760789] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 641.760789] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.760789] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 641.760789] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.760789] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 641.760789] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.760789] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 641.760789] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.760789] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 641.760789] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.760789] env[61857]: ERROR nova.compute.manager raise self.value [ 641.760789] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.760789] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 641.760789] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.760789] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 641.761489] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.761489] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 641.761489] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c382853b-2a41-415a-918c-d07424264041, please check neutron logs for more information. [ 641.761489] env[61857]: ERROR nova.compute.manager [ 641.761489] env[61857]: Traceback (most recent call last): [ 641.761489] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 641.761489] env[61857]: listener.cb(fileno) [ 641.761489] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.761489] env[61857]: result = function(*args, **kwargs) [ 641.761489] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.761489] env[61857]: return func(*args, **kwargs) [ 641.761489] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.761489] env[61857]: raise e [ 641.761489] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.761489] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 641.761489] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.761489] env[61857]: created_port_ids = self._update_ports_for_instance( [ 641.761489] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.761489] env[61857]: with excutils.save_and_reraise_exception(): [ 641.761489] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.761489] env[61857]: self.force_reraise() [ 641.761489] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.761489] env[61857]: raise self.value [ 641.761489] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.761489] env[61857]: updated_port = self._update_port( [ 641.761489] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.761489] env[61857]: _ensure_no_port_binding_failure(port) [ 641.761489] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.761489] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 641.762471] env[61857]: nova.exception.PortBindingFailed: Binding failed for port c382853b-2a41-415a-918c-d07424264041, please check neutron logs for more information. [ 641.762471] env[61857]: Removing descriptor: 16 [ 641.819257] env[61857]: DEBUG nova.scheduler.client.report [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.831613] env[61857]: DEBUG nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 641.860784] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:38:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='40ab1924-8be9-40ef-ac38-6e56405739b6',id=32,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-333534802',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 641.860784] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 641.860784] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.861061] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 641.861183] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.861345] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 641.861533] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 641.861864] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 641.861864] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 641.862425] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 641.862425] env[61857]: DEBUG nova.virt.hardware [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.863174] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf0c406-86e3-40c9-8748-3be0f2e5e34f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.871888] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad8b04d-245f-490e-af80-9238c20acfcb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.888018] env[61857]: ERROR nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c382853b-2a41-415a-918c-d07424264041, please check neutron logs for more information. [ 641.888018] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Traceback (most recent call last): [ 641.888018] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 641.888018] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] yield resources [ 641.888018] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.888018] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] self.driver.spawn(context, instance, image_meta, [ 641.888018] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 641.888018] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.888018] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.888018] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] vm_ref = self.build_virtual_machine(instance, [ 641.888018] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] for vif in network_info: [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] return self._sync_wrapper(fn, *args, **kwargs) [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] self.wait() [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] self[:] = self._gt.wait() [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] return self._exit_event.wait() [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 641.888513] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] current.throw(*self._exc) [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] result = function(*args, **kwargs) [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] return func(*args, **kwargs) [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] raise e [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] nwinfo = self.network_api.allocate_for_instance( [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] created_port_ids = self._update_ports_for_instance( [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] with excutils.save_and_reraise_exception(): [ 641.888885] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] self.force_reraise() [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] raise self.value [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] updated_port = self._update_port( [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] _ensure_no_port_binding_failure(port) [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] raise exception.PortBindingFailed(port_id=port['id']) [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] nova.exception.PortBindingFailed: Binding failed for port c382853b-2a41-415a-918c-d07424264041, please check neutron logs for more information. [ 641.889298] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] [ 641.889298] env[61857]: INFO nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Terminating instance [ 641.889647] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquiring lock "refresh_cache-ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.039676] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.128158] env[61857]: DEBUG nova.network.neutron [req-a7332305-b8b2-49ab-a209-b34d302b7e0c req-9deef5ce-fb87-4f72-83d3-33336d3884dc service nova] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.202134] env[61857]: DEBUG nova.network.neutron [req-a7332305-b8b2-49ab-a209-b34d302b7e0c req-9deef5ce-fb87-4f72-83d3-33336d3884dc service nova] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.210368] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Acquiring lock "c4883f84-e62b-4b59-9484-270d82dc34e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.210612] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Lock "c4883f84-e62b-4b59-9484-270d82dc34e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.324630] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.325166] env[61857]: DEBUG nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 642.327811] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.355s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.705299] env[61857]: DEBUG oslo_concurrency.lockutils [req-a7332305-b8b2-49ab-a209-b34d302b7e0c req-9deef5ce-fb87-4f72-83d3-33336d3884dc service nova] Releasing lock "refresh_cache-ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.705740] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquired lock "refresh_cache-ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.705927] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 642.833453] env[61857]: DEBUG nova.compute.utils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 642.839920] env[61857]: DEBUG nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 642.839920] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 642.887481] env[61857]: DEBUG nova.policy [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c21b2b2013cd40a69cc7a958a821e377', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ff67efd6404381a52a4001ec774a72', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.235257] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.240755] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Successfully created port: 9a41d152-65a1-4a6b-8c9e-0020354a0092 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.292181] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.339752] env[61857]: DEBUG nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 643.366721] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c1526323-a057-4a84-aa32-05cd802ebf9d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 643.366877] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance dafceac5-1eea-4103-8a48-ca6c5136390e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 643.367011] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 406f6726-4009-49bd-a37b-91707ece1932 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 643.367137] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 3375702e-36fd-43d4-95da-e3aafb96a32f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 643.367252] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 1492bf97-9e03-40eb-ae6c-041f2998124e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 643.367392] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 643.367552] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 7e466be0-f291-43c3-a79c-fb3871f06c23 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 643.633571] env[61857]: DEBUG nova.compute.manager [req-c3b032ec-390a-40ad-bff3-a5a4ae02b3cc req-d7967470-2315-4091-b19d-896d73f295a8 service nova] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Received event network-vif-deleted-c382853b-2a41-415a-918c-d07424264041 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 643.794741] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Releasing lock "refresh_cache-ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.795205] env[61857]: DEBUG nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.795421] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 643.795767] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f8cb903e-e78d-4b6c-95eb-8d9512d69903 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.809790] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f189bd-1116-4aae-bc30-e0a01ca51097 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.841890] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75 could not be found. [ 643.842163] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 643.842356] env[61857]: INFO nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Took 0.05 seconds to destroy the instance on the hypervisor. [ 643.842643] env[61857]: DEBUG oslo.service.loopingcall [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.842879] env[61857]: DEBUG nova.compute.manager [-] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.842976] env[61857]: DEBUG nova.network.neutron [-] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 643.861815] env[61857]: DEBUG nova.network.neutron [-] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.870229] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 249ffa46-c5c7-4669-84a5-c02e0799dc59 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.250695] env[61857]: ERROR nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9a41d152-65a1-4a6b-8c9e-0020354a0092, please check neutron logs for more information. [ 644.250695] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 644.250695] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.250695] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 644.250695] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.250695] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 644.250695] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.250695] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 644.250695] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.250695] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 644.250695] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.250695] env[61857]: ERROR nova.compute.manager raise self.value [ 644.250695] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.250695] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 644.250695] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.250695] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 644.251204] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.251204] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 644.251204] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9a41d152-65a1-4a6b-8c9e-0020354a0092, please check neutron logs for more information. [ 644.251204] env[61857]: ERROR nova.compute.manager [ 644.251204] env[61857]: Traceback (most recent call last): [ 644.251204] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 644.251204] env[61857]: listener.cb(fileno) [ 644.251204] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.251204] env[61857]: result = function(*args, **kwargs) [ 644.251204] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.251204] env[61857]: return func(*args, **kwargs) [ 644.251204] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.251204] env[61857]: raise e [ 644.251204] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.251204] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 644.251204] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.251204] env[61857]: created_port_ids = self._update_ports_for_instance( [ 644.251204] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.251204] env[61857]: with excutils.save_and_reraise_exception(): [ 644.251204] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.251204] env[61857]: self.force_reraise() [ 644.251204] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.251204] env[61857]: raise self.value [ 644.251204] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.251204] env[61857]: updated_port = self._update_port( [ 644.251204] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.251204] env[61857]: _ensure_no_port_binding_failure(port) [ 644.251204] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.251204] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 644.251902] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 9a41d152-65a1-4a6b-8c9e-0020354a0092, please check neutron logs for more information. [ 644.251902] env[61857]: Removing descriptor: 16 [ 644.352836] env[61857]: DEBUG nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 644.368400] env[61857]: DEBUG nova.network.neutron [-] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.374144] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 95fbef26-78a7-4449-ba06-ffd517fb1032 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.381936] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 644.382199] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 644.382432] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 644.382624] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 644.382814] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 644.383018] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 644.383270] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 644.383413] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 644.383578] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 644.383739] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 644.383907] env[61857]: DEBUG nova.virt.hardware [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 644.384803] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7cfbf84-9d6c-4a0e-9266-a90b66399347 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.394995] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38abbce6-51cd-462d-a977-e9fc5829538c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.409569] env[61857]: ERROR nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9a41d152-65a1-4a6b-8c9e-0020354a0092, please check neutron logs for more information. [ 644.409569] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Traceback (most recent call last): [ 644.409569] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 644.409569] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] yield resources [ 644.409569] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.409569] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] self.driver.spawn(context, instance, image_meta, [ 644.409569] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 644.409569] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.409569] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.409569] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] vm_ref = self.build_virtual_machine(instance, [ 644.409569] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] for vif in network_info: [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] return self._sync_wrapper(fn, *args, **kwargs) [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] self.wait() [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] self[:] = self._gt.wait() [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] return self._exit_event.wait() [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 644.409954] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] current.throw(*self._exc) [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] result = function(*args, **kwargs) [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] return func(*args, **kwargs) [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] raise e [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] nwinfo = self.network_api.allocate_for_instance( [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] created_port_ids = self._update_ports_for_instance( [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] with excutils.save_and_reraise_exception(): [ 644.410356] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] self.force_reraise() [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] raise self.value [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] updated_port = self._update_port( [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] _ensure_no_port_binding_failure(port) [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] raise exception.PortBindingFailed(port_id=port['id']) [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] nova.exception.PortBindingFailed: Binding failed for port 9a41d152-65a1-4a6b-8c9e-0020354a0092, please check neutron logs for more information. [ 644.410740] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] [ 644.410740] env[61857]: INFO nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Terminating instance [ 644.413056] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "refresh_cache-7e466be0-f291-43c3-a79c-fb3871f06c23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.413056] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquired lock "refresh_cache-7e466be0-f291-43c3-a79c-fb3871f06c23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.413216] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 644.870723] env[61857]: INFO nova.compute.manager [-] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Took 1.03 seconds to deallocate network for instance. [ 644.873330] env[61857]: DEBUG nova.compute.claims [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.874055] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.877350] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 4ede4082-93b0-4cac-9e28-5448f04cba9b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.930536] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.035559] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.381074] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 51c54fe3-c0e2-4151-9626-cfb10cd7996e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.538522] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Releasing lock "refresh_cache-7e466be0-f291-43c3-a79c-fb3871f06c23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.538975] env[61857]: DEBUG nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 645.539191] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 645.539491] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4adeed9-35aa-4b55-a50a-b7e3ec8a34e4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.549235] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b3bdd0-c54b-444b-af6c-cc68c3067aaf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.574242] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7e466be0-f291-43c3-a79c-fb3871f06c23 could not be found. [ 645.574492] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 645.574688] env[61857]: INFO nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Took 0.04 seconds to destroy the instance on the hypervisor. [ 645.574926] env[61857]: DEBUG oslo.service.loopingcall [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 645.575162] env[61857]: DEBUG nova.compute.manager [-] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.575258] env[61857]: DEBUG nova.network.neutron [-] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 645.589758] env[61857]: DEBUG nova.network.neutron [-] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.659544] env[61857]: DEBUG nova.compute.manager [req-653d8da7-3820-44cd-b87b-e728819998f0 req-8ef35dc2-ecdb-437b-980a-fb687616d230 service nova] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Received event network-changed-9a41d152-65a1-4a6b-8c9e-0020354a0092 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 645.660088] env[61857]: DEBUG nova.compute.manager [req-653d8da7-3820-44cd-b87b-e728819998f0 req-8ef35dc2-ecdb-437b-980a-fb687616d230 service nova] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Refreshing instance network info cache due to event network-changed-9a41d152-65a1-4a6b-8c9e-0020354a0092. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 645.660313] env[61857]: DEBUG oslo_concurrency.lockutils [req-653d8da7-3820-44cd-b87b-e728819998f0 req-8ef35dc2-ecdb-437b-980a-fb687616d230 service nova] Acquiring lock "refresh_cache-7e466be0-f291-43c3-a79c-fb3871f06c23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.660545] env[61857]: DEBUG oslo_concurrency.lockutils [req-653d8da7-3820-44cd-b87b-e728819998f0 req-8ef35dc2-ecdb-437b-980a-fb687616d230 service nova] Acquired lock "refresh_cache-7e466be0-f291-43c3-a79c-fb3871f06c23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.661134] env[61857]: DEBUG nova.network.neutron [req-653d8da7-3820-44cd-b87b-e728819998f0 req-8ef35dc2-ecdb-437b-980a-fb687616d230 service nova] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Refreshing network info cache for port 9a41d152-65a1-4a6b-8c9e-0020354a0092 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 645.884063] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 8ee206cf-c906-4ac0-8ba7-d864f5f538db has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.096021] env[61857]: DEBUG nova.network.neutron [-] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.177451] env[61857]: DEBUG nova.network.neutron [req-653d8da7-3820-44cd-b87b-e728819998f0 req-8ef35dc2-ecdb-437b-980a-fb687616d230 service nova] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.259570] env[61857]: DEBUG nova.network.neutron [req-653d8da7-3820-44cd-b87b-e728819998f0 req-8ef35dc2-ecdb-437b-980a-fb687616d230 service nova] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.387142] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ca3d3ad9-39a0-467b-812f-21b7303b8bc8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.598154] env[61857]: INFO nova.compute.manager [-] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Took 1.02 seconds to deallocate network for instance. [ 646.600172] env[61857]: DEBUG nova.compute.claims [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 646.600420] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.762233] env[61857]: DEBUG oslo_concurrency.lockutils [req-653d8da7-3820-44cd-b87b-e728819998f0 req-8ef35dc2-ecdb-437b-980a-fb687616d230 service nova] Releasing lock "refresh_cache-7e466be0-f291-43c3-a79c-fb3871f06c23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.762595] env[61857]: DEBUG nova.compute.manager [req-653d8da7-3820-44cd-b87b-e728819998f0 req-8ef35dc2-ecdb-437b-980a-fb687616d230 service nova] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Received event network-vif-deleted-9a41d152-65a1-4a6b-8c9e-0020354a0092 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 646.889825] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 9edf9e8c-5421-4018-afee-59e10cd335ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.395882] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e52972ca-877f-45b8-b79f-c834246a8857 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.898620] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 952d0a1c-3e94-42dc-9bc4-8063275afc82 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.401715] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 768e0e7d-7ebe-4d7a-a971-4fb463731050 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.904851] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance b3fd3980-efe4-49a0-b656-32f1543e657c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.408289] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.911928] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 2c187f64-8a36-4dfd-94e3-8ea944dbac24 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.415054] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 6304714b-8988-4798-990f-d3e19774eb8f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.918917] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 11adaee5-8e56-4679-8373-5d5690a44ca9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.421947] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e88aed62-6561-4e29-b041-46b19cc5ef63 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.924874] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c4f70321-f612-416e-b209-eb974dab9d49 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.428098] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 18d9307d-6a20-47b4-b4e9-176f9a8c33cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.930451] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance d49008e6-7789-4785-8eea-9b3fa36c4ccb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.433816] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 3cacef7b-dd24-4d33-9500-bbff03e342bb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.936818] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e286e6d0-25db-4d86-b9a2-3538bc3f624a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.440405] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 7a800abe-ea0a-4077-a5da-dd60eab917e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.943721] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c4883f84-e62b-4b59-9484-270d82dc34e0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.944011] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 654.944174] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 655.268307] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f694e2a-565b-4d3e-9462-538f1380b17c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.276273] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ba5e0d-f00c-4c01-be7b-93f001b92819 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.306620] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8009e03-2c57-4a6f-a1c7-6ef8a7291f28 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.314263] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a00fd20-178d-497c-b408-f6cd40127264 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.327923] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.831498] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.338452] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 656.338772] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.011s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.339066] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.314s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.340600] env[61857]: INFO nova.compute.claims [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.669140] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a096039-faed-4a32-8dc1-e527c737ab43 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.676759] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642125b7-cde3-46a5-b0aa-e5ba2611936e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.708016] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181dc1e7-0097-4812-9d60-cd5a38a94f6a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.716417] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8631ed6f-7fd7-4f76-a742-838dad1d0cbd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.732645] env[61857]: DEBUG nova.compute.provider_tree [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.236188] env[61857]: DEBUG nova.scheduler.client.report [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.742025] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.742321] env[61857]: DEBUG nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 658.744770] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.403s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.249425] env[61857]: DEBUG nova.compute.utils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.254302] env[61857]: DEBUG nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 659.254562] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 659.299904] env[61857]: DEBUG nova.policy [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac8387df3290404283263440672de653', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3bbcf7c2937040e1906e2273a07b671b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 659.580926] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Successfully created port: 8262c9df-5f69-417b-b554-64d69d14d52d {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 659.698387] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ecada9e-1aa8-401f-a611-3fd7e7cde5e8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.711311] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e28695-bcb2-4e5f-b196-90c2ffd777cc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.743774] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eaca6ad-f0b6-4e4b-9572-9f42f0edcfbb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.753875] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c0e59a-c291-4f34-b923-51d072ea5462 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.758382] env[61857]: DEBUG nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 659.771511] env[61857]: DEBUG nova.compute.provider_tree [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.274821] env[61857]: DEBUG nova.scheduler.client.report [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.427249] env[61857]: DEBUG nova.compute.manager [req-843c25a4-e8ec-4098-a62b-c88059d661c0 req-f5bdba70-8c62-4e58-811e-de150f2bedc4 service nova] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Received event network-changed-8262c9df-5f69-417b-b554-64d69d14d52d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 660.427463] env[61857]: DEBUG nova.compute.manager [req-843c25a4-e8ec-4098-a62b-c88059d661c0 req-f5bdba70-8c62-4e58-811e-de150f2bedc4 service nova] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Refreshing instance network info cache due to event network-changed-8262c9df-5f69-417b-b554-64d69d14d52d. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 660.427665] env[61857]: DEBUG oslo_concurrency.lockutils [req-843c25a4-e8ec-4098-a62b-c88059d661c0 req-f5bdba70-8c62-4e58-811e-de150f2bedc4 service nova] Acquiring lock "refresh_cache-249ffa46-c5c7-4669-84a5-c02e0799dc59" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.427810] env[61857]: DEBUG oslo_concurrency.lockutils [req-843c25a4-e8ec-4098-a62b-c88059d661c0 req-f5bdba70-8c62-4e58-811e-de150f2bedc4 service nova] Acquired lock "refresh_cache-249ffa46-c5c7-4669-84a5-c02e0799dc59" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.427972] env[61857]: DEBUG nova.network.neutron [req-843c25a4-e8ec-4098-a62b-c88059d661c0 req-f5bdba70-8c62-4e58-811e-de150f2bedc4 service nova] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Refreshing network info cache for port 8262c9df-5f69-417b-b554-64d69d14d52d {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 660.613364] env[61857]: ERROR nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8262c9df-5f69-417b-b554-64d69d14d52d, please check neutron logs for more information. [ 660.613364] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.613364] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.613364] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.613364] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.613364] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.613364] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.613364] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.613364] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.613364] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 660.613364] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.613364] env[61857]: ERROR nova.compute.manager raise self.value [ 660.613364] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.613364] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.613364] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.613364] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.613865] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.613865] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.613865] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8262c9df-5f69-417b-b554-64d69d14d52d, please check neutron logs for more information. [ 660.613865] env[61857]: ERROR nova.compute.manager [ 660.613865] env[61857]: Traceback (most recent call last): [ 660.613865] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.613865] env[61857]: listener.cb(fileno) [ 660.613865] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.613865] env[61857]: result = function(*args, **kwargs) [ 660.613865] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.613865] env[61857]: return func(*args, **kwargs) [ 660.613865] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.613865] env[61857]: raise e [ 660.613865] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.613865] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 660.613865] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.613865] env[61857]: created_port_ids = self._update_ports_for_instance( [ 660.613865] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.613865] env[61857]: with excutils.save_and_reraise_exception(): [ 660.613865] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.613865] env[61857]: self.force_reraise() [ 660.613865] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.613865] env[61857]: raise self.value [ 660.613865] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.613865] env[61857]: updated_port = self._update_port( [ 660.613865] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.613865] env[61857]: _ensure_no_port_binding_failure(port) [ 660.613865] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.613865] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.614781] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 8262c9df-5f69-417b-b554-64d69d14d52d, please check neutron logs for more information. [ 660.614781] env[61857]: Removing descriptor: 16 [ 660.768570] env[61857]: DEBUG nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 660.780767] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.036s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.781206] env[61857]: ERROR nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa, please check neutron logs for more information. [ 660.781206] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Traceback (most recent call last): [ 660.781206] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.781206] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] self.driver.spawn(context, instance, image_meta, [ 660.781206] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 660.781206] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.781206] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.781206] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] vm_ref = self.build_virtual_machine(instance, [ 660.781206] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.781206] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.781206] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] for vif in network_info: [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] return self._sync_wrapper(fn, *args, **kwargs) [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] self.wait() [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] self[:] = self._gt.wait() [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] return self._exit_event.wait() [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] result = hub.switch() [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.781574] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] return self.greenlet.switch() [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] result = function(*args, **kwargs) [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] return func(*args, **kwargs) [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] raise e [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] nwinfo = self.network_api.allocate_for_instance( [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] created_port_ids = self._update_ports_for_instance( [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] with excutils.save_and_reraise_exception(): [ 660.781946] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] self.force_reraise() [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] raise self.value [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] updated_port = self._update_port( [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] _ensure_no_port_binding_failure(port) [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] raise exception.PortBindingFailed(port_id=port['id']) [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] nova.exception.PortBindingFailed: Binding failed for port 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa, please check neutron logs for more information. [ 660.782325] env[61857]: ERROR nova.compute.manager [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] [ 660.782655] env[61857]: DEBUG nova.compute.utils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Binding failed for port 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.783038] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.577s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.784479] env[61857]: INFO nova.compute.claims [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.792977] env[61857]: DEBUG nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Build of instance c1526323-a057-4a84-aa32-05cd802ebf9d was re-scheduled: Binding failed for port 9bb00c41-d1cd-4f3b-ac26-1dd7231b8efa, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 660.793687] env[61857]: DEBUG nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 660.794826] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Acquiring lock "refresh_cache-c1526323-a057-4a84-aa32-05cd802ebf9d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.794826] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Acquired lock "refresh_cache-c1526323-a057-4a84-aa32-05cd802ebf9d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.794826] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 660.797253] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 660.797474] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 660.797633] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 660.797816] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 660.797962] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 660.798161] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 660.798379] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 660.798538] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 660.798703] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 660.798867] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 660.799054] env[61857]: DEBUG nova.virt.hardware [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 660.799907] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7da76d-b4c5-4bc8-be8d-7b15d356ef6c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.812026] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8be8bf-18e3-4f2c-a796-fb32d94105c0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.827960] env[61857]: ERROR nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8262c9df-5f69-417b-b554-64d69d14d52d, please check neutron logs for more information. [ 660.827960] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Traceback (most recent call last): [ 660.827960] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 660.827960] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] yield resources [ 660.827960] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.827960] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] self.driver.spawn(context, instance, image_meta, [ 660.827960] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 660.827960] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.827960] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.827960] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] vm_ref = self.build_virtual_machine(instance, [ 660.827960] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] for vif in network_info: [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] return self._sync_wrapper(fn, *args, **kwargs) [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] self.wait() [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] self[:] = self._gt.wait() [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] return self._exit_event.wait() [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 660.828329] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] current.throw(*self._exc) [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] result = function(*args, **kwargs) [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] return func(*args, **kwargs) [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] raise e [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] nwinfo = self.network_api.allocate_for_instance( [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] created_port_ids = self._update_ports_for_instance( [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] with excutils.save_and_reraise_exception(): [ 660.828655] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] self.force_reraise() [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] raise self.value [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] updated_port = self._update_port( [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] _ensure_no_port_binding_failure(port) [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] raise exception.PortBindingFailed(port_id=port['id']) [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] nova.exception.PortBindingFailed: Binding failed for port 8262c9df-5f69-417b-b554-64d69d14d52d, please check neutron logs for more information. [ 660.829201] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] [ 660.829201] env[61857]: INFO nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Terminating instance [ 660.830359] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.832522] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-249ffa46-c5c7-4669-84a5-c02e0799dc59" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.925902] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.944762] env[61857]: DEBUG nova.network.neutron [req-843c25a4-e8ec-4098-a62b-c88059d661c0 req-f5bdba70-8c62-4e58-811e-de150f2bedc4 service nova] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.014311] env[61857]: DEBUG nova.network.neutron [req-843c25a4-e8ec-4098-a62b-c88059d661c0 req-f5bdba70-8c62-4e58-811e-de150f2bedc4 service nova] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.428901] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Releasing lock "refresh_cache-c1526323-a057-4a84-aa32-05cd802ebf9d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.429190] env[61857]: DEBUG nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 661.429380] env[61857]: DEBUG nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.429554] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 661.446938] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.517686] env[61857]: DEBUG oslo_concurrency.lockutils [req-843c25a4-e8ec-4098-a62b-c88059d661c0 req-f5bdba70-8c62-4e58-811e-de150f2bedc4 service nova] Releasing lock "refresh_cache-249ffa46-c5c7-4669-84a5-c02e0799dc59" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.518213] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-249ffa46-c5c7-4669-84a5-c02e0799dc59" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.518421] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.949830] env[61857]: DEBUG nova.network.neutron [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.037328] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.123865] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.132708] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59b5be9-7db5-42d9-9e9c-6ed43b07ce8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.141728] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18188e4-d6b5-4fe9-a3dd-13173ed61ae7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.172756] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7f0521-f1e4-4b0d-9606-643d26a18eed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.180234] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f5eab8-de09-4899-9ee9-e53c7545fea6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.194965] env[61857]: DEBUG nova.compute.provider_tree [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.451918] env[61857]: INFO nova.compute.manager [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] [instance: c1526323-a057-4a84-aa32-05cd802ebf9d] Took 1.02 seconds to deallocate network for instance. [ 662.464894] env[61857]: DEBUG nova.compute.manager [req-b2dec327-2591-4037-ae49-fae841c28c9b req-673ac6f6-9cb5-4f25-be54-79802c229bdf service nova] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Received event network-vif-deleted-8262c9df-5f69-417b-b554-64d69d14d52d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 662.626963] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-249ffa46-c5c7-4669-84a5-c02e0799dc59" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.627432] env[61857]: DEBUG nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 662.627653] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 662.627951] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08e7ab4f-1732-4b1c-aab8-b7c246a4bcad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.637798] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda60a17-7826-4711-9278-be14cc9df124 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.661784] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 249ffa46-c5c7-4669-84a5-c02e0799dc59 could not be found. [ 662.662021] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 662.662215] env[61857]: INFO nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Took 0.03 seconds to destroy the instance on the hypervisor. [ 662.662524] env[61857]: DEBUG oslo.service.loopingcall [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.662774] env[61857]: DEBUG nova.compute.manager [-] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.662914] env[61857]: DEBUG nova.network.neutron [-] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.679988] env[61857]: DEBUG nova.network.neutron [-] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.697893] env[61857]: DEBUG nova.scheduler.client.report [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.183925] env[61857]: DEBUG nova.network.neutron [-] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.202293] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.202834] env[61857]: DEBUG nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 663.205291] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.758s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.481554] env[61857]: INFO nova.scheduler.client.report [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Deleted allocations for instance c1526323-a057-4a84-aa32-05cd802ebf9d [ 663.685348] env[61857]: INFO nova.compute.manager [-] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Took 1.02 seconds to deallocate network for instance. [ 663.687747] env[61857]: DEBUG nova.compute.claims [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 663.687922] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.709301] env[61857]: DEBUG nova.compute.utils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.710858] env[61857]: DEBUG nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.711031] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 663.759769] env[61857]: DEBUG nova.policy [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '543d5489afd54191b511c4d21551aa6d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ccfd5b9778a84bd995ac3252cff8537d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.989559] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dba3cbf1-a0b1-45a3-91a0-19d21c3dd639 tempest-ImagesNegativeTestJSON-943531206 tempest-ImagesNegativeTestJSON-943531206-project-member] Lock "c1526323-a057-4a84-aa32-05cd802ebf9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.437s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.086278] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Successfully created port: ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.118817] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13831305-45ce-470c-a1a8-453deccef94e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.129184] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72a3b81b-8be8-4c3d-be8e-5caff5b65f61 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.160753] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aff7255-95bc-46fd-b89d-91fabe85c040 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.170301] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a40107d4-f0a8-4805-826d-ce4b75f04313 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.185111] env[61857]: DEBUG nova.compute.provider_tree [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.216510] env[61857]: DEBUG nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 664.492502] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 664.688492] env[61857]: DEBUG nova.scheduler.client.report [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.870279] env[61857]: DEBUG nova.compute.manager [req-5f01bb39-98f3-452d-bc72-e1dcbb8227e0 req-c91438fe-1a9f-4b66-9e9c-b0131f64768a service nova] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Received event network-changed-ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 664.870507] env[61857]: DEBUG nova.compute.manager [req-5f01bb39-98f3-452d-bc72-e1dcbb8227e0 req-c91438fe-1a9f-4b66-9e9c-b0131f64768a service nova] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Refreshing instance network info cache due to event network-changed-ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 664.870777] env[61857]: DEBUG oslo_concurrency.lockutils [req-5f01bb39-98f3-452d-bc72-e1dcbb8227e0 req-c91438fe-1a9f-4b66-9e9c-b0131f64768a service nova] Acquiring lock "refresh_cache-95fbef26-78a7-4449-ba06-ffd517fb1032" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.870988] env[61857]: DEBUG oslo_concurrency.lockutils [req-5f01bb39-98f3-452d-bc72-e1dcbb8227e0 req-c91438fe-1a9f-4b66-9e9c-b0131f64768a service nova] Acquired lock "refresh_cache-95fbef26-78a7-4449-ba06-ffd517fb1032" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.871364] env[61857]: DEBUG nova.network.neutron [req-5f01bb39-98f3-452d-bc72-e1dcbb8227e0 req-c91438fe-1a9f-4b66-9e9c-b0131f64768a service nova] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Refreshing network info cache for port ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 665.021379] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.095601] env[61857]: ERROR nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c, please check neutron logs for more information. [ 665.095601] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 665.095601] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.095601] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 665.095601] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.095601] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 665.095601] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.095601] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 665.095601] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.095601] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 665.095601] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.095601] env[61857]: ERROR nova.compute.manager raise self.value [ 665.095601] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.095601] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 665.095601] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.095601] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 665.096297] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.096297] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 665.096297] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c, please check neutron logs for more information. [ 665.096297] env[61857]: ERROR nova.compute.manager [ 665.096297] env[61857]: Traceback (most recent call last): [ 665.096297] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 665.096297] env[61857]: listener.cb(fileno) [ 665.096297] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.096297] env[61857]: result = function(*args, **kwargs) [ 665.096297] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.096297] env[61857]: return func(*args, **kwargs) [ 665.096297] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.096297] env[61857]: raise e [ 665.096297] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.096297] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 665.096297] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.096297] env[61857]: created_port_ids = self._update_ports_for_instance( [ 665.096297] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.096297] env[61857]: with excutils.save_and_reraise_exception(): [ 665.096297] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.096297] env[61857]: self.force_reraise() [ 665.096297] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.096297] env[61857]: raise self.value [ 665.096297] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.096297] env[61857]: updated_port = self._update_port( [ 665.096297] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.096297] env[61857]: _ensure_no_port_binding_failure(port) [ 665.096297] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.096297] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 665.097375] env[61857]: nova.exception.PortBindingFailed: Binding failed for port ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c, please check neutron logs for more information. [ 665.097375] env[61857]: Removing descriptor: 16 [ 665.193383] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.194363] env[61857]: ERROR nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a1f70ff1-faa9-40bd-9540-e9edfaabcdd9, please check neutron logs for more information. [ 665.194363] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Traceback (most recent call last): [ 665.194363] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.194363] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] self.driver.spawn(context, instance, image_meta, [ 665.194363] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 665.194363] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.194363] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.194363] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] vm_ref = self.build_virtual_machine(instance, [ 665.194363] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.194363] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.194363] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] for vif in network_info: [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] return self._sync_wrapper(fn, *args, **kwargs) [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] self.wait() [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] self[:] = self._gt.wait() [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] return self._exit_event.wait() [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] result = hub.switch() [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.194656] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] return self.greenlet.switch() [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] result = function(*args, **kwargs) [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] return func(*args, **kwargs) [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] raise e [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] nwinfo = self.network_api.allocate_for_instance( [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] created_port_ids = self._update_ports_for_instance( [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] with excutils.save_and_reraise_exception(): [ 665.194959] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] self.force_reraise() [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] raise self.value [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] updated_port = self._update_port( [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] _ensure_no_port_binding_failure(port) [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] raise exception.PortBindingFailed(port_id=port['id']) [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] nova.exception.PortBindingFailed: Binding failed for port a1f70ff1-faa9-40bd-9540-e9edfaabcdd9, please check neutron logs for more information. [ 665.195277] env[61857]: ERROR nova.compute.manager [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] [ 665.195542] env[61857]: DEBUG nova.compute.utils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Binding failed for port a1f70ff1-faa9-40bd-9540-e9edfaabcdd9, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.197234] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.158s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.201888] env[61857]: DEBUG nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Build of instance dafceac5-1eea-4103-8a48-ca6c5136390e was re-scheduled: Binding failed for port a1f70ff1-faa9-40bd-9540-e9edfaabcdd9, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 665.202524] env[61857]: DEBUG nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 665.203088] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquiring lock "refresh_cache-dafceac5-1eea-4103-8a48-ca6c5136390e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.203088] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Acquired lock "refresh_cache-dafceac5-1eea-4103-8a48-ca6c5136390e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.203330] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 665.233225] env[61857]: DEBUG nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 665.267621] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 665.267908] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 665.268546] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.268546] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 665.268546] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.268546] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 665.268858] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 665.269021] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 665.269195] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 665.269365] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 665.269536] env[61857]: DEBUG nova.virt.hardware [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 665.270442] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec72e28a-f13f-476b-a6e4-25cea41165d4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.281395] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd2dcea-2452-4c9d-8804-fb0ed905fc99 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.297431] env[61857]: ERROR nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c, please check neutron logs for more information. [ 665.297431] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Traceback (most recent call last): [ 665.297431] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.297431] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] yield resources [ 665.297431] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.297431] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] self.driver.spawn(context, instance, image_meta, [ 665.297431] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 665.297431] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.297431] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.297431] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] vm_ref = self.build_virtual_machine(instance, [ 665.297431] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] for vif in network_info: [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] return self._sync_wrapper(fn, *args, **kwargs) [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] self.wait() [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] self[:] = self._gt.wait() [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] return self._exit_event.wait() [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 665.297757] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] current.throw(*self._exc) [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] result = function(*args, **kwargs) [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] return func(*args, **kwargs) [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] raise e [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] nwinfo = self.network_api.allocate_for_instance( [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] created_port_ids = self._update_ports_for_instance( [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] with excutils.save_and_reraise_exception(): [ 665.298225] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] self.force_reraise() [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] raise self.value [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] updated_port = self._update_port( [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] _ensure_no_port_binding_failure(port) [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] raise exception.PortBindingFailed(port_id=port['id']) [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] nova.exception.PortBindingFailed: Binding failed for port ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c, please check neutron logs for more information. [ 665.298555] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] [ 665.298555] env[61857]: INFO nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Terminating instance [ 665.299741] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Acquiring lock "refresh_cache-95fbef26-78a7-4449-ba06-ffd517fb1032" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.388671] env[61857]: DEBUG nova.network.neutron [req-5f01bb39-98f3-452d-bc72-e1dcbb8227e0 req-c91438fe-1a9f-4b66-9e9c-b0131f64768a service nova] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.473735] env[61857]: DEBUG nova.network.neutron [req-5f01bb39-98f3-452d-bc72-e1dcbb8227e0 req-c91438fe-1a9f-4b66-9e9c-b0131f64768a service nova] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.721561] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.788226] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.976489] env[61857]: DEBUG oslo_concurrency.lockutils [req-5f01bb39-98f3-452d-bc72-e1dcbb8227e0 req-c91438fe-1a9f-4b66-9e9c-b0131f64768a service nova] Releasing lock "refresh_cache-95fbef26-78a7-4449-ba06-ffd517fb1032" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.976887] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Acquired lock "refresh_cache-95fbef26-78a7-4449-ba06-ffd517fb1032" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.977089] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.082781] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56a4cda-8bbc-4cbd-8356-af4bf64d4de6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.090460] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6379b34-c33d-4511-88ef-ff95f80d72af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.121813] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1635f4ab-9477-4274-9c49-fbe2078ac359 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.129140] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4170faeb-d40f-4271-a2ca-31d87ef61591 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.143553] env[61857]: DEBUG nova.compute.provider_tree [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.293364] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Releasing lock "refresh_cache-dafceac5-1eea-4103-8a48-ca6c5136390e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.296230] env[61857]: DEBUG nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 666.296230] env[61857]: DEBUG nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.296230] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 666.314633] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.512825] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.647949] env[61857]: DEBUG nova.scheduler.client.report [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.652421] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.819531] env[61857]: DEBUG nova.network.neutron [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.897010] env[61857]: DEBUG nova.compute.manager [req-dbeba786-2e30-48d6-b789-045413132d5d req-da13b416-bf9d-4ffe-95e4-239abf39e5e6 service nova] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Received event network-vif-deleted-ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 667.154566] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.957s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.155244] env[61857]: ERROR nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c612b995-5fce-4231-a737-bdc5b3b06635, please check neutron logs for more information. [ 667.155244] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] Traceback (most recent call last): [ 667.155244] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 667.155244] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] self.driver.spawn(context, instance, image_meta, [ 667.155244] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 667.155244] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.155244] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.155244] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] vm_ref = self.build_virtual_machine(instance, [ 667.155244] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.155244] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.155244] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] for vif in network_info: [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] return self._sync_wrapper(fn, *args, **kwargs) [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] self.wait() [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] self[:] = self._gt.wait() [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] return self._exit_event.wait() [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] result = hub.switch() [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.155588] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] return self.greenlet.switch() [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] result = function(*args, **kwargs) [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] return func(*args, **kwargs) [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] raise e [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] nwinfo = self.network_api.allocate_for_instance( [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] created_port_ids = self._update_ports_for_instance( [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] with excutils.save_and_reraise_exception(): [ 667.155928] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] self.force_reraise() [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] raise self.value [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] updated_port = self._update_port( [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] _ensure_no_port_binding_failure(port) [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] raise exception.PortBindingFailed(port_id=port['id']) [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] nova.exception.PortBindingFailed: Binding failed for port c612b995-5fce-4231-a737-bdc5b3b06635, please check neutron logs for more information. [ 667.156275] env[61857]: ERROR nova.compute.manager [instance: 406f6726-4009-49bd-a37b-91707ece1932] [ 667.156568] env[61857]: DEBUG nova.compute.utils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Binding failed for port c612b995-5fce-4231-a737-bdc5b3b06635, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 667.157278] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Releasing lock "refresh_cache-95fbef26-78a7-4449-ba06-ffd517fb1032" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.157700] env[61857]: DEBUG nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.157943] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 667.158299] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.125s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.162091] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e636ad3-841e-487b-a171-1d1ac90f460e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.164409] env[61857]: DEBUG nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Build of instance 406f6726-4009-49bd-a37b-91707ece1932 was re-scheduled: Binding failed for port c612b995-5fce-4231-a737-bdc5b3b06635, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 667.164849] env[61857]: DEBUG nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 667.165079] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquiring lock "refresh_cache-406f6726-4009-49bd-a37b-91707ece1932" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.165230] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Acquired lock "refresh_cache-406f6726-4009-49bd-a37b-91707ece1932" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.165430] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 667.173300] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08fb282c-a450-4546-98dc-d66cdf81869c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.197663] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 95fbef26-78a7-4449-ba06-ffd517fb1032 could not be found. [ 667.197900] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 667.198147] env[61857]: INFO nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Took 0.04 seconds to destroy the instance on the hypervisor. [ 667.198410] env[61857]: DEBUG oslo.service.loopingcall [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.198635] env[61857]: DEBUG nova.compute.manager [-] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.198733] env[61857]: DEBUG nova.network.neutron [-] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.215047] env[61857]: DEBUG nova.network.neutron [-] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.322643] env[61857]: INFO nova.compute.manager [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] [instance: dafceac5-1eea-4103-8a48-ca6c5136390e] Took 1.03 seconds to deallocate network for instance. [ 667.687138] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.718350] env[61857]: DEBUG nova.network.neutron [-] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.861765] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.141820] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f629bc47-7c52-4b30-a72b-830c48adb4a5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.151668] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfbb2da5-f940-4b29-b2e9-5e0bd983e670 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.183240] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053ac901-e07b-4a49-ae1b-2f56b03b7c5a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.190772] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9c0a22-1660-465b-a012-d3cbe3a49c78 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.208021] env[61857]: DEBUG nova.compute.provider_tree [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.226465] env[61857]: INFO nova.compute.manager [-] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Took 1.03 seconds to deallocate network for instance. [ 668.228842] env[61857]: DEBUG nova.compute.claims [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 668.229053] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.361551] env[61857]: INFO nova.scheduler.client.report [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Deleted allocations for instance dafceac5-1eea-4103-8a48-ca6c5136390e [ 668.368194] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Releasing lock "refresh_cache-406f6726-4009-49bd-a37b-91707ece1932" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.371723] env[61857]: DEBUG nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 668.371819] env[61857]: DEBUG nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.371998] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 668.400454] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.712019] env[61857]: DEBUG nova.scheduler.client.report [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.878315] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3a8539d0-d392-4d2d-8472-5361a93c689b tempest-DeleteServersAdminTestJSON-2051289013 tempest-DeleteServersAdminTestJSON-2051289013-project-member] Lock "dafceac5-1eea-4103-8a48-ca6c5136390e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.767s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.905070] env[61857]: DEBUG nova.network.neutron [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.215174] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.057s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.215831] env[61857]: ERROR nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ae0205a7-5adf-4aae-a937-72a665c82f24, please check neutron logs for more information. [ 669.215831] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Traceback (most recent call last): [ 669.215831] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.215831] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] self.driver.spawn(context, instance, image_meta, [ 669.215831] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.215831] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.215831] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.215831] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] vm_ref = self.build_virtual_machine(instance, [ 669.215831] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.215831] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.215831] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] for vif in network_info: [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] return self._sync_wrapper(fn, *args, **kwargs) [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] self.wait() [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] self[:] = self._gt.wait() [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] return self._exit_event.wait() [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] result = hub.switch() [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.216135] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] return self.greenlet.switch() [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] result = function(*args, **kwargs) [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] return func(*args, **kwargs) [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] raise e [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] nwinfo = self.network_api.allocate_for_instance( [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] created_port_ids = self._update_ports_for_instance( [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] with excutils.save_and_reraise_exception(): [ 669.216519] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] self.force_reraise() [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] raise self.value [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] updated_port = self._update_port( [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] _ensure_no_port_binding_failure(port) [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] raise exception.PortBindingFailed(port_id=port['id']) [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] nova.exception.PortBindingFailed: Binding failed for port ae0205a7-5adf-4aae-a937-72a665c82f24, please check neutron logs for more information. [ 669.217272] env[61857]: ERROR nova.compute.manager [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] [ 669.218509] env[61857]: DEBUG nova.compute.utils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Binding failed for port ae0205a7-5adf-4aae-a937-72a665c82f24, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 669.218509] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.691s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.221251] env[61857]: DEBUG nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Build of instance 3375702e-36fd-43d4-95da-e3aafb96a32f was re-scheduled: Binding failed for port ae0205a7-5adf-4aae-a937-72a665c82f24, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 669.221703] env[61857]: DEBUG nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 669.221930] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Acquiring lock "refresh_cache-3375702e-36fd-43d4-95da-e3aafb96a32f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.225164] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Acquired lock "refresh_cache-3375702e-36fd-43d4-95da-e3aafb96a32f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.225364] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 669.382216] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 669.411216] env[61857]: INFO nova.compute.manager [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] [instance: 406f6726-4009-49bd-a37b-91707ece1932] Took 1.04 seconds to deallocate network for instance. [ 669.750988] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.890716] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.926907] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.222057] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b35ee0-d039-4b73-9093-c421dc33befa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.230620] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d710113e-e8fd-4315-b088-9455dabfaf99 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.261772] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ac9173-a074-44c0-98b9-8c60b8cb32a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.269502] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41a23ac-f6e2-441b-aa2f-2d096bf63a5e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.283189] env[61857]: DEBUG nova.compute.provider_tree [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.396575] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Releasing lock "refresh_cache-3375702e-36fd-43d4-95da-e3aafb96a32f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.396870] env[61857]: DEBUG nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 670.397038] env[61857]: DEBUG nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.397209] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 670.421068] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.457817] env[61857]: INFO nova.scheduler.client.report [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Deleted allocations for instance 406f6726-4009-49bd-a37b-91707ece1932 [ 670.570272] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "781a2790-e317-46fe-9be8-40c9e5f1f771" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.570521] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.787155] env[61857]: DEBUG nova.scheduler.client.report [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.921862] env[61857]: DEBUG nova.network.neutron [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.965694] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b868dbd5-aa9c-4f06-b4af-eb0fb6b15500 tempest-ListServerFiltersTestJSON-990577062 tempest-ListServerFiltersTestJSON-990577062-project-member] Lock "406f6726-4009-49bd-a37b-91707ece1932" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.262s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.291384] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.073s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.292063] env[61857]: ERROR nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 20789189-3d73-49fc-a0ed-04d7cdd3cb78, please check neutron logs for more information. [ 671.292063] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Traceback (most recent call last): [ 671.292063] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.292063] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] self.driver.spawn(context, instance, image_meta, [ 671.292063] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 671.292063] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.292063] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.292063] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] vm_ref = self.build_virtual_machine(instance, [ 671.292063] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.292063] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.292063] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] for vif in network_info: [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] return self._sync_wrapper(fn, *args, **kwargs) [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] self.wait() [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] self[:] = self._gt.wait() [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] return self._exit_event.wait() [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] result = hub.switch() [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.292549] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] return self.greenlet.switch() [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] result = function(*args, **kwargs) [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] return func(*args, **kwargs) [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] raise e [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] nwinfo = self.network_api.allocate_for_instance( [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] created_port_ids = self._update_ports_for_instance( [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] with excutils.save_and_reraise_exception(): [ 671.292856] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] self.force_reraise() [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] raise self.value [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] updated_port = self._update_port( [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] _ensure_no_port_binding_failure(port) [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] raise exception.PortBindingFailed(port_id=port['id']) [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] nova.exception.PortBindingFailed: Binding failed for port 20789189-3d73-49fc-a0ed-04d7cdd3cb78, please check neutron logs for more information. [ 671.293195] env[61857]: ERROR nova.compute.manager [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] [ 671.293458] env[61857]: DEBUG nova.compute.utils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Binding failed for port 20789189-3d73-49fc-a0ed-04d7cdd3cb78, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 671.296970] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.254s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.299753] env[61857]: INFO nova.compute.claims [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.304863] env[61857]: DEBUG nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Build of instance 1492bf97-9e03-40eb-ae6c-041f2998124e was re-scheduled: Binding failed for port 20789189-3d73-49fc-a0ed-04d7cdd3cb78, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 671.305383] env[61857]: DEBUG nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 671.305640] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Acquiring lock "refresh_cache-1492bf97-9e03-40eb-ae6c-041f2998124e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.305701] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Acquired lock "refresh_cache-1492bf97-9e03-40eb-ae6c-041f2998124e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.305864] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.426194] env[61857]: INFO nova.compute.manager [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] [instance: 3375702e-36fd-43d4-95da-e3aafb96a32f] Took 1.03 seconds to deallocate network for instance. [ 671.469231] env[61857]: DEBUG nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 671.832439] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.929837] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.994367] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.435532] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Releasing lock "refresh_cache-1492bf97-9e03-40eb-ae6c-041f2998124e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.435787] env[61857]: DEBUG nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.435997] env[61857]: DEBUG nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.436121] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.464416] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.466277] env[61857]: INFO nova.scheduler.client.report [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Deleted allocations for instance 3375702e-36fd-43d4-95da-e3aafb96a32f [ 672.728423] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b4146d-37b4-4506-85e3-267ac3bd18f3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.738498] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bd97c5-0481-4382-a7fc-ead2f068ce50 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.772016] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5d43a7-ef9b-4ec2-8e75-56c7fff2ab40 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.779565] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf608da-22bb-4098-8c84-5f0adb5f647e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.796261] env[61857]: DEBUG nova.compute.provider_tree [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.974616] env[61857]: DEBUG nova.network.neutron [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.979650] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6f167c6a-9733-4627-8474-d4a629bbf6b6 tempest-ServerRescueTestJSON-1273550563 tempest-ServerRescueTestJSON-1273550563-project-member] Lock "3375702e-36fd-43d4-95da-e3aafb96a32f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.462s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.299987] env[61857]: DEBUG nova.scheduler.client.report [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.477388] env[61857]: INFO nova.compute.manager [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] [instance: 1492bf97-9e03-40eb-ae6c-041f2998124e] Took 1.04 seconds to deallocate network for instance. [ 673.485596] env[61857]: DEBUG nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.804824] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.805351] env[61857]: DEBUG nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 673.808485] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.935s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.010335] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.313821] env[61857]: DEBUG nova.compute.utils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 674.319403] env[61857]: DEBUG nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 674.321428] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 674.394063] env[61857]: DEBUG nova.policy [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '572389a0121b4046ab4fb0c39c647dc6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06cebaf6852e4a3fb1bf415204a62421', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 674.518434] env[61857]: INFO nova.scheduler.client.report [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Deleted allocations for instance 1492bf97-9e03-40eb-ae6c-041f2998124e [ 674.725795] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4fbe7f-548d-4ceb-a6c1-7184bdce8795 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.737444] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27126b6c-a339-447f-a40a-7b5c382b7fc9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.767428] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d58654-5c66-47b9-9040-9111811ed701 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.775395] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55be6a83-deb6-4736-8f51-f5793d5f8e41 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.789746] env[61857]: DEBUG nova.compute.provider_tree [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.819703] env[61857]: DEBUG nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.905054] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Successfully created port: 9069d1ca-2999-4d67-a1ca-eaa5b68f051b {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 675.031161] env[61857]: DEBUG oslo_concurrency.lockutils [None req-96c65c0e-de9d-43ec-940b-bdd765ab6090 tempest-ServersTestManualDisk-325337118 tempest-ServersTestManualDisk-325337118-project-member] Lock "1492bf97-9e03-40eb-ae6c-041f2998124e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.002s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.293919] env[61857]: DEBUG nova.scheduler.client.report [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.539212] env[61857]: DEBUG nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.800081] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.800800] env[61857]: ERROR nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c382853b-2a41-415a-918c-d07424264041, please check neutron logs for more information. [ 675.800800] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Traceback (most recent call last): [ 675.800800] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.800800] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] self.driver.spawn(context, instance, image_meta, [ 675.800800] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.800800] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.800800] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.800800] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] vm_ref = self.build_virtual_machine(instance, [ 675.800800] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.800800] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.800800] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] for vif in network_info: [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] return self._sync_wrapper(fn, *args, **kwargs) [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] self.wait() [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] self[:] = self._gt.wait() [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] return self._exit_event.wait() [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] current.throw(*self._exc) [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.801126] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] result = function(*args, **kwargs) [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] return func(*args, **kwargs) [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] raise e [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] nwinfo = self.network_api.allocate_for_instance( [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] created_port_ids = self._update_ports_for_instance( [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] with excutils.save_and_reraise_exception(): [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] self.force_reraise() [ 675.801524] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.801898] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] raise self.value [ 675.801898] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.801898] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] updated_port = self._update_port( [ 675.801898] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.801898] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] _ensure_no_port_binding_failure(port) [ 675.801898] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.801898] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] raise exception.PortBindingFailed(port_id=port['id']) [ 675.801898] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] nova.exception.PortBindingFailed: Binding failed for port c382853b-2a41-415a-918c-d07424264041, please check neutron logs for more information. [ 675.801898] env[61857]: ERROR nova.compute.manager [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] [ 675.801898] env[61857]: DEBUG nova.compute.utils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Binding failed for port c382853b-2a41-415a-918c-d07424264041, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.805164] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.203s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.808835] env[61857]: DEBUG nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Build of instance ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75 was re-scheduled: Binding failed for port c382853b-2a41-415a-918c-d07424264041, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 675.809326] env[61857]: DEBUG nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 675.809590] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquiring lock "refresh_cache-ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.809799] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Acquired lock "refresh_cache-ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.809967] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 675.831491] env[61857]: DEBUG nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 675.872732] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 675.872732] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 675.872732] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 675.872986] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 675.872986] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 675.874163] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 675.874437] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 675.874610] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 675.874778] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 675.874941] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 675.876678] env[61857]: DEBUG nova.virt.hardware [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.877512] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6055d702-1af3-4170-8d83-a6755bb7a98a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.887398] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a2ac95-6d11-4e8e-8cfa-46e7cbd01229 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.084133] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.347365] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.522115] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.682603] env[61857]: ERROR nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9069d1ca-2999-4d67-a1ca-eaa5b68f051b, please check neutron logs for more information. [ 676.682603] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 676.682603] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.682603] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 676.682603] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.682603] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 676.682603] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.682603] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 676.682603] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.682603] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 676.682603] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.682603] env[61857]: ERROR nova.compute.manager raise self.value [ 676.682603] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.682603] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 676.682603] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.682603] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 676.683333] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.683333] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 676.683333] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9069d1ca-2999-4d67-a1ca-eaa5b68f051b, please check neutron logs for more information. [ 676.683333] env[61857]: ERROR nova.compute.manager [ 676.683333] env[61857]: Traceback (most recent call last): [ 676.683333] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 676.683333] env[61857]: listener.cb(fileno) [ 676.683333] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.683333] env[61857]: result = function(*args, **kwargs) [ 676.683333] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.683333] env[61857]: return func(*args, **kwargs) [ 676.683333] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.683333] env[61857]: raise e [ 676.683333] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.683333] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 676.683333] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.683333] env[61857]: created_port_ids = self._update_ports_for_instance( [ 676.683333] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.683333] env[61857]: with excutils.save_and_reraise_exception(): [ 676.683333] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.683333] env[61857]: self.force_reraise() [ 676.683333] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.683333] env[61857]: raise self.value [ 676.683333] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.683333] env[61857]: updated_port = self._update_port( [ 676.683333] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.683333] env[61857]: _ensure_no_port_binding_failure(port) [ 676.683333] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.683333] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 676.685081] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 9069d1ca-2999-4d67-a1ca-eaa5b68f051b, please check neutron logs for more information. [ 676.685081] env[61857]: Removing descriptor: 16 [ 676.685081] env[61857]: ERROR nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9069d1ca-2999-4d67-a1ca-eaa5b68f051b, please check neutron logs for more information. [ 676.685081] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Traceback (most recent call last): [ 676.685081] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 676.685081] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] yield resources [ 676.685081] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.685081] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] self.driver.spawn(context, instance, image_meta, [ 676.685081] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 676.685081] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.685081] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.685081] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] vm_ref = self.build_virtual_machine(instance, [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] for vif in network_info: [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] return self._sync_wrapper(fn, *args, **kwargs) [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] self.wait() [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] self[:] = self._gt.wait() [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] return self._exit_event.wait() [ 676.685391] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] result = hub.switch() [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] return self.greenlet.switch() [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] result = function(*args, **kwargs) [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] return func(*args, **kwargs) [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] raise e [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] nwinfo = self.network_api.allocate_for_instance( [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.685738] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] created_port_ids = self._update_ports_for_instance( [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] with excutils.save_and_reraise_exception(): [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] self.force_reraise() [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] raise self.value [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] updated_port = self._update_port( [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] _ensure_no_port_binding_failure(port) [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.686128] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] raise exception.PortBindingFailed(port_id=port['id']) [ 676.686517] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] nova.exception.PortBindingFailed: Binding failed for port 9069d1ca-2999-4d67-a1ca-eaa5b68f051b, please check neutron logs for more information. [ 676.686517] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] [ 676.686517] env[61857]: INFO nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Terminating instance [ 676.689743] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Acquiring lock "refresh_cache-4ede4082-93b0-4cac-9e28-5448f04cba9b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.689743] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Acquired lock "refresh_cache-4ede4082-93b0-4cac-9e28-5448f04cba9b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.689966] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.748273] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37c626b-4178-4078-9bc4-510dedd65bb3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.759883] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96de9877-b45d-4852-8883-efa8ed9e2afb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.793248] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4676eb9-0693-4d15-8c3c-366764d1ffe3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.800608] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0775c652-c32f-49bc-9619-8d615fb0a0a2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.813966] env[61857]: DEBUG nova.compute.provider_tree [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.844912] env[61857]: DEBUG nova.compute.manager [req-781b2ee3-e6ad-474f-a977-10aa69b234a0 req-8e2ad5bd-81de-40dc-a4f0-93a75d678f87 service nova] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Received event network-changed-9069d1ca-2999-4d67-a1ca-eaa5b68f051b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 676.844912] env[61857]: DEBUG nova.compute.manager [req-781b2ee3-e6ad-474f-a977-10aa69b234a0 req-8e2ad5bd-81de-40dc-a4f0-93a75d678f87 service nova] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Refreshing instance network info cache due to event network-changed-9069d1ca-2999-4d67-a1ca-eaa5b68f051b. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 676.844912] env[61857]: DEBUG oslo_concurrency.lockutils [req-781b2ee3-e6ad-474f-a977-10aa69b234a0 req-8e2ad5bd-81de-40dc-a4f0-93a75d678f87 service nova] Acquiring lock "refresh_cache-4ede4082-93b0-4cac-9e28-5448f04cba9b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.026637] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Releasing lock "refresh_cache-ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.026637] env[61857]: DEBUG nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 677.026637] env[61857]: DEBUG nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.026637] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.048975] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.212503] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.320725] env[61857]: DEBUG nova.scheduler.client.report [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.406792] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.553585] env[61857]: DEBUG nova.network.neutron [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.826566] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.023s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.828276] env[61857]: ERROR nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9a41d152-65a1-4a6b-8c9e-0020354a0092, please check neutron logs for more information. [ 677.828276] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Traceback (most recent call last): [ 677.828276] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.828276] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] self.driver.spawn(context, instance, image_meta, [ 677.828276] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.828276] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.828276] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.828276] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] vm_ref = self.build_virtual_machine(instance, [ 677.828276] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.828276] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.828276] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] for vif in network_info: [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] return self._sync_wrapper(fn, *args, **kwargs) [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] self.wait() [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] self[:] = self._gt.wait() [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] return self._exit_event.wait() [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] current.throw(*self._exc) [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.828610] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] result = function(*args, **kwargs) [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] return func(*args, **kwargs) [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] raise e [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] nwinfo = self.network_api.allocate_for_instance( [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] created_port_ids = self._update_ports_for_instance( [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] with excutils.save_and_reraise_exception(): [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] self.force_reraise() [ 677.828919] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.829238] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] raise self.value [ 677.829238] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.829238] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] updated_port = self._update_port( [ 677.829238] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.829238] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] _ensure_no_port_binding_failure(port) [ 677.829238] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.829238] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] raise exception.PortBindingFailed(port_id=port['id']) [ 677.829238] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] nova.exception.PortBindingFailed: Binding failed for port 9a41d152-65a1-4a6b-8c9e-0020354a0092, please check neutron logs for more information. [ 677.829238] env[61857]: ERROR nova.compute.manager [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] [ 677.829451] env[61857]: DEBUG nova.compute.utils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Binding failed for port 9a41d152-65a1-4a6b-8c9e-0020354a0092, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.830920] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.143s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.839692] env[61857]: DEBUG nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Build of instance 7e466be0-f291-43c3-a79c-fb3871f06c23 was re-scheduled: Binding failed for port 9a41d152-65a1-4a6b-8c9e-0020354a0092, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.839692] env[61857]: DEBUG nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.839692] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "refresh_cache-7e466be0-f291-43c3-a79c-fb3871f06c23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.839692] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquired lock "refresh_cache-7e466be0-f291-43c3-a79c-fb3871f06c23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.839927] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.910050] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Releasing lock "refresh_cache-4ede4082-93b0-4cac-9e28-5448f04cba9b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.910487] env[61857]: DEBUG nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 677.913548] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 677.913548] env[61857]: DEBUG oslo_concurrency.lockutils [req-781b2ee3-e6ad-474f-a977-10aa69b234a0 req-8e2ad5bd-81de-40dc-a4f0-93a75d678f87 service nova] Acquired lock "refresh_cache-4ede4082-93b0-4cac-9e28-5448f04cba9b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.913548] env[61857]: DEBUG nova.network.neutron [req-781b2ee3-e6ad-474f-a977-10aa69b234a0 req-8e2ad5bd-81de-40dc-a4f0-93a75d678f87 service nova] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Refreshing network info cache for port 9069d1ca-2999-4d67-a1ca-eaa5b68f051b {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 677.913548] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9dc2ff3f-8a39-4d31-b999-19df27bdd163 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.926585] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599d2995-32c9-40b6-aefd-eccf06be9ae9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.951320] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4ede4082-93b0-4cac-9e28-5448f04cba9b could not be found. [ 677.951578] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 677.951759] env[61857]: INFO nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 677.952020] env[61857]: DEBUG oslo.service.loopingcall [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 677.952238] env[61857]: DEBUG nova.compute.manager [-] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.952332] env[61857]: DEBUG nova.network.neutron [-] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.978371] env[61857]: DEBUG nova.network.neutron [-] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.058034] env[61857]: INFO nova.compute.manager [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] [instance: ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75] Took 1.03 seconds to deallocate network for instance. [ 678.377185] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.446345] env[61857]: DEBUG nova.network.neutron [req-781b2ee3-e6ad-474f-a977-10aa69b234a0 req-8e2ad5bd-81de-40dc-a4f0-93a75d678f87 service nova] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.480076] env[61857]: DEBUG nova.network.neutron [-] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.533951] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.642337] env[61857]: DEBUG nova.network.neutron [req-781b2ee3-e6ad-474f-a977-10aa69b234a0 req-8e2ad5bd-81de-40dc-a4f0-93a75d678f87 service nova] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.793574] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284ba13a-fb9c-4f15-af01-28f947bc3fe5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.800289] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d0625d-88dd-460d-964f-7611af4905b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.831809] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7068f70f-043e-4429-a478-b27a124e9a27 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.843665] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc2ccbb-1a15-4144-8dc7-88643c6b4c46 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.858761] env[61857]: DEBUG nova.compute.provider_tree [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.982852] env[61857]: INFO nova.compute.manager [-] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Took 1.03 seconds to deallocate network for instance. [ 678.985220] env[61857]: DEBUG nova.compute.claims [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 678.985499] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.039599] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Releasing lock "refresh_cache-7e466be0-f291-43c3-a79c-fb3871f06c23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.039599] env[61857]: DEBUG nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 679.039599] env[61857]: DEBUG nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.039599] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 679.059476] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.078352] env[61857]: DEBUG nova.compute.manager [req-bf1adedf-48f8-4dc6-897b-21171f18019c req-161e78f5-7f6e-459d-9815-4f2b97715134 service nova] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Received event network-vif-deleted-9069d1ca-2999-4d67-a1ca-eaa5b68f051b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 679.097019] env[61857]: INFO nova.scheduler.client.report [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Deleted allocations for instance ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75 [ 679.146298] env[61857]: DEBUG oslo_concurrency.lockutils [req-781b2ee3-e6ad-474f-a977-10aa69b234a0 req-8e2ad5bd-81de-40dc-a4f0-93a75d678f87 service nova] Releasing lock "refresh_cache-4ede4082-93b0-4cac-9e28-5448f04cba9b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.365173] env[61857]: DEBUG nova.scheduler.client.report [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.561043] env[61857]: DEBUG nova.network.neutron [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.602382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d4e7554-1853-4e08-b783-1053d56f5f36 tempest-MigrationsAdminTest-1918901483 tempest-MigrationsAdminTest-1918901483-project-member] Lock "ad6986fe-0d48-49e2-a03b-e0ae1e6f3b75" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.297s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.868109] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.037s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.868786] env[61857]: ERROR nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8262c9df-5f69-417b-b554-64d69d14d52d, please check neutron logs for more information. [ 679.868786] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Traceback (most recent call last): [ 679.868786] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.868786] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] self.driver.spawn(context, instance, image_meta, [ 679.868786] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 679.868786] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.868786] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.868786] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] vm_ref = self.build_virtual_machine(instance, [ 679.868786] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.868786] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.868786] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] for vif in network_info: [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] return self._sync_wrapper(fn, *args, **kwargs) [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] self.wait() [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] self[:] = self._gt.wait() [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] return self._exit_event.wait() [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] current.throw(*self._exc) [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.869109] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] result = function(*args, **kwargs) [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] return func(*args, **kwargs) [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] raise e [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] nwinfo = self.network_api.allocate_for_instance( [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] created_port_ids = self._update_ports_for_instance( [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] with excutils.save_and_reraise_exception(): [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] self.force_reraise() [ 679.869457] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.869883] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] raise self.value [ 679.869883] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.869883] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] updated_port = self._update_port( [ 679.869883] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.869883] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] _ensure_no_port_binding_failure(port) [ 679.869883] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.869883] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] raise exception.PortBindingFailed(port_id=port['id']) [ 679.869883] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] nova.exception.PortBindingFailed: Binding failed for port 8262c9df-5f69-417b-b554-64d69d14d52d, please check neutron logs for more information. [ 679.869883] env[61857]: ERROR nova.compute.manager [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] [ 679.869883] env[61857]: DEBUG nova.compute.utils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Binding failed for port 8262c9df-5f69-417b-b554-64d69d14d52d, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.870901] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.850s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.872854] env[61857]: INFO nova.compute.claims [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.875513] env[61857]: DEBUG nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Build of instance 249ffa46-c5c7-4669-84a5-c02e0799dc59 was re-scheduled: Binding failed for port 8262c9df-5f69-417b-b554-64d69d14d52d, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.876023] env[61857]: DEBUG nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.876289] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-249ffa46-c5c7-4669-84a5-c02e0799dc59" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.876467] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-249ffa46-c5c7-4669-84a5-c02e0799dc59" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.876704] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 680.063678] env[61857]: INFO nova.compute.manager [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 7e466be0-f291-43c3-a79c-fb3871f06c23] Took 1.03 seconds to deallocate network for instance. [ 680.107519] env[61857]: DEBUG nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 680.211685] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "71920a62-090d-4df1-937e-87df3b043e28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.211870] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.403893] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.514512] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.633302] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.017330] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-249ffa46-c5c7-4669-84a5-c02e0799dc59" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.017620] env[61857]: DEBUG nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.017750] env[61857]: DEBUG nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.017919] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.043355] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.098683] env[61857]: INFO nova.scheduler.client.report [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Deleted allocations for instance 7e466be0-f291-43c3-a79c-fb3871f06c23 [ 681.307250] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72dd56b-3b37-4af7-a629-d2e355560e6d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.315186] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531f1776-f759-4bb5-8831-a9fab4047059 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.346782] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376fa716-ccaa-4ad4-bc77-acca40eda81d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.359203] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca646ab-16e7-4f80-8506-24c7bcecd1a3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.370956] env[61857]: DEBUG nova.compute.provider_tree [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.545527] env[61857]: DEBUG nova.network.neutron [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.616030] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fe2bfda-74db-4f8b-b5f9-fdd6338f9229 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "7e466be0-f291-43c3-a79c-fb3871f06c23" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.555s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.874765] env[61857]: DEBUG nova.scheduler.client.report [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.048805] env[61857]: INFO nova.compute.manager [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 249ffa46-c5c7-4669-84a5-c02e0799dc59] Took 1.03 seconds to deallocate network for instance. [ 682.119824] env[61857]: DEBUG nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 682.386981] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.389932] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 682.391382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.162s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.658266] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.901130] env[61857]: DEBUG nova.compute.utils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.905914] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 682.907341] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 682.968043] env[61857]: DEBUG nova.policy [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a3a9944299d4574bf46978852f11d92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b21cc6a46cc04f20b942603dc80ec6c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 683.098957] env[61857]: INFO nova.scheduler.client.report [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleted allocations for instance 249ffa46-c5c7-4669-84a5-c02e0799dc59 [ 683.400273] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f77382-68a3-4287-a728-117cc4265cda {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.406482] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 683.418499] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92430bb3-7f4d-4021-ad6c-ae0641394cb3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.457135] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588178de-b4a8-4dbc-bda2-f4bac245b080 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.465589] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59e35af-476f-4f08-9d88-f5813a3786e8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.483216] env[61857]: DEBUG nova.compute.provider_tree [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.618640] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7ae6df1-4402-4cd6-84f3-4ec9586829c7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "249ffa46-c5c7-4669-84a5-c02e0799dc59" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.673s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.692900] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Successfully created port: fe95c043-c85d-470c-ac22-755174721da1 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.985924] env[61857]: DEBUG nova.scheduler.client.report [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.123338] env[61857]: DEBUG nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 684.429026] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 684.469700] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.469939] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.470474] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.470723] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.472378] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.472556] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.472778] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.473718] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.473916] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.474105] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.474284] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.475559] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25655f8d-4c98-4119-bb7d-07d221bcbcfe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.485340] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27fb9192-e347-4104-890b-6a3599477eb6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.493655] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.102s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.494268] env[61857]: ERROR nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c, please check neutron logs for more information. [ 684.494268] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Traceback (most recent call last): [ 684.494268] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.494268] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] self.driver.spawn(context, instance, image_meta, [ 684.494268] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 684.494268] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.494268] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.494268] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] vm_ref = self.build_virtual_machine(instance, [ 684.494268] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.494268] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.494268] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] for vif in network_info: [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] return self._sync_wrapper(fn, *args, **kwargs) [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] self.wait() [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] self[:] = self._gt.wait() [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] return self._exit_event.wait() [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] current.throw(*self._exc) [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.494583] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] result = function(*args, **kwargs) [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] return func(*args, **kwargs) [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] raise e [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] nwinfo = self.network_api.allocate_for_instance( [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] created_port_ids = self._update_ports_for_instance( [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] with excutils.save_and_reraise_exception(): [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] self.force_reraise() [ 684.494959] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.495344] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] raise self.value [ 684.495344] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.495344] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] updated_port = self._update_port( [ 684.495344] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.495344] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] _ensure_no_port_binding_failure(port) [ 684.495344] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.495344] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] raise exception.PortBindingFailed(port_id=port['id']) [ 684.495344] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] nova.exception.PortBindingFailed: Binding failed for port ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c, please check neutron logs for more information. [ 684.495344] env[61857]: ERROR nova.compute.manager [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] [ 684.495344] env[61857]: DEBUG nova.compute.utils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Binding failed for port ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.496253] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.570s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.498255] env[61857]: INFO nova.compute.claims [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.502532] env[61857]: DEBUG nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Build of instance 95fbef26-78a7-4449-ba06-ffd517fb1032 was re-scheduled: Binding failed for port ba2a8ba3-1c5a-4660-a197-ecfc8c29dd3c, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.503901] env[61857]: DEBUG nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.503901] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Acquiring lock "refresh_cache-95fbef26-78a7-4449-ba06-ffd517fb1032" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.503901] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Acquired lock "refresh_cache-95fbef26-78a7-4449-ba06-ffd517fb1032" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.503901] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 684.609197] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.609506] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.649703] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.054926] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.238613] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.545169] env[61857]: ERROR nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fe95c043-c85d-470c-ac22-755174721da1, please check neutron logs for more information. [ 685.545169] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 685.545169] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.545169] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 685.545169] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 685.545169] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 685.545169] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 685.545169] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 685.545169] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.545169] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 685.545169] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.545169] env[61857]: ERROR nova.compute.manager raise self.value [ 685.545169] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 685.545169] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 685.545169] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.545169] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 685.545912] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.545912] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 685.545912] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fe95c043-c85d-470c-ac22-755174721da1, please check neutron logs for more information. [ 685.545912] env[61857]: ERROR nova.compute.manager [ 685.545912] env[61857]: Traceback (most recent call last): [ 685.545912] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 685.545912] env[61857]: listener.cb(fileno) [ 685.545912] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.545912] env[61857]: result = function(*args, **kwargs) [ 685.545912] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.545912] env[61857]: return func(*args, **kwargs) [ 685.545912] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.545912] env[61857]: raise e [ 685.545912] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.545912] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 685.545912] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 685.545912] env[61857]: created_port_ids = self._update_ports_for_instance( [ 685.545912] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 685.545912] env[61857]: with excutils.save_and_reraise_exception(): [ 685.545912] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.545912] env[61857]: self.force_reraise() [ 685.545912] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.545912] env[61857]: raise self.value [ 685.545912] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 685.545912] env[61857]: updated_port = self._update_port( [ 685.545912] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.545912] env[61857]: _ensure_no_port_binding_failure(port) [ 685.545912] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.545912] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 685.547080] env[61857]: nova.exception.PortBindingFailed: Binding failed for port fe95c043-c85d-470c-ac22-755174721da1, please check neutron logs for more information. [ 685.547080] env[61857]: Removing descriptor: 17 [ 685.547080] env[61857]: ERROR nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fe95c043-c85d-470c-ac22-755174721da1, please check neutron logs for more information. [ 685.547080] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Traceback (most recent call last): [ 685.547080] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 685.547080] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] yield resources [ 685.547080] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.547080] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] self.driver.spawn(context, instance, image_meta, [ 685.547080] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 685.547080] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.547080] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.547080] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] vm_ref = self.build_virtual_machine(instance, [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] for vif in network_info: [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] return self._sync_wrapper(fn, *args, **kwargs) [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] self.wait() [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] self[:] = self._gt.wait() [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] return self._exit_event.wait() [ 685.547574] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] result = hub.switch() [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] return self.greenlet.switch() [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] result = function(*args, **kwargs) [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] return func(*args, **kwargs) [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] raise e [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] nwinfo = self.network_api.allocate_for_instance( [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 685.548146] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] created_port_ids = self._update_ports_for_instance( [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] with excutils.save_and_reraise_exception(): [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] self.force_reraise() [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] raise self.value [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] updated_port = self._update_port( [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] _ensure_no_port_binding_failure(port) [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.548783] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] raise exception.PortBindingFailed(port_id=port['id']) [ 685.550448] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] nova.exception.PortBindingFailed: Binding failed for port fe95c043-c85d-470c-ac22-755174721da1, please check neutron logs for more information. [ 685.550448] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] [ 685.550448] env[61857]: INFO nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Terminating instance [ 685.550448] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "refresh_cache-51c54fe3-c0e2-4151-9626-cfb10cd7996e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.550448] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired lock "refresh_cache-51c54fe3-c0e2-4151-9626-cfb10cd7996e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.550448] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.742853] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Releasing lock "refresh_cache-95fbef26-78a7-4449-ba06-ffd517fb1032" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.743984] env[61857]: DEBUG nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.744207] env[61857]: DEBUG nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.744385] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 685.778288] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.827366] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.827593] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.828851] env[61857]: DEBUG nova.compute.manager [req-eb36dd9c-5b74-4668-8fd2-f0311f69ca73 req-e894dd0d-adb6-4fad-ab7e-28f027b3c843 service nova] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Received event network-changed-fe95c043-c85d-470c-ac22-755174721da1 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 685.828851] env[61857]: DEBUG nova.compute.manager [req-eb36dd9c-5b74-4668-8fd2-f0311f69ca73 req-e894dd0d-adb6-4fad-ab7e-28f027b3c843 service nova] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Refreshing instance network info cache due to event network-changed-fe95c043-c85d-470c-ac22-755174721da1. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 685.828967] env[61857]: DEBUG oslo_concurrency.lockutils [req-eb36dd9c-5b74-4668-8fd2-f0311f69ca73 req-e894dd0d-adb6-4fad-ab7e-28f027b3c843 service nova] Acquiring lock "refresh_cache-51c54fe3-c0e2-4151-9626-cfb10cd7996e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.045057] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3370f03-415e-44cb-b48e-bf0a52b2f5cf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.053942] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc8090d-fd88-4e55-b8cb-bc1a6c517cf3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.065487] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "e91e99b9-1cd3-4345-af09-f14af4df1214" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.065807] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "e91e99b9-1cd3-4345-af09-f14af4df1214" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.096187] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.098798] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2b9582-f506-49f0-af56-54188ffc5660 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.107412] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0401ce2-e7c0-471a-b723-d4210db7b139 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.127258] env[61857]: DEBUG nova.compute.provider_tree [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.232717] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.280559] env[61857]: DEBUG nova.network.neutron [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.630377] env[61857]: DEBUG nova.scheduler.client.report [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.737199] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Releasing lock "refresh_cache-51c54fe3-c0e2-4151-9626-cfb10cd7996e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.737199] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 686.737199] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 686.737199] env[61857]: DEBUG oslo_concurrency.lockutils [req-eb36dd9c-5b74-4668-8fd2-f0311f69ca73 req-e894dd0d-adb6-4fad-ab7e-28f027b3c843 service nova] Acquired lock "refresh_cache-51c54fe3-c0e2-4151-9626-cfb10cd7996e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.737352] env[61857]: DEBUG nova.network.neutron [req-eb36dd9c-5b74-4668-8fd2-f0311f69ca73 req-e894dd0d-adb6-4fad-ab7e-28f027b3c843 service nova] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Refreshing network info cache for port fe95c043-c85d-470c-ac22-755174721da1 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 686.738368] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-360a5930-4945-4753-b1f7-db95d821bafb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.748529] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88721494-950a-4c1c-97a4-0c8f30c94da8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.774400] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 51c54fe3-c0e2-4151-9626-cfb10cd7996e could not be found. [ 686.774678] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 686.774755] env[61857]: INFO nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 686.775339] env[61857]: DEBUG oslo.service.loopingcall [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.775522] env[61857]: DEBUG nova.compute.manager [-] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.775595] env[61857]: DEBUG nova.network.neutron [-] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 686.788319] env[61857]: INFO nova.compute.manager [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] [instance: 95fbef26-78a7-4449-ba06-ffd517fb1032] Took 1.04 seconds to deallocate network for instance. [ 686.795395] env[61857]: DEBUG nova.network.neutron [-] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.945040] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "30d905ed-831d-44ae-807c-062de9a7e9fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.949144] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.136282] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.640s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.136913] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 687.140500] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.146s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.141398] env[61857]: INFO nova.compute.claims [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.258193] env[61857]: DEBUG nova.network.neutron [req-eb36dd9c-5b74-4668-8fd2-f0311f69ca73 req-e894dd0d-adb6-4fad-ab7e-28f027b3c843 service nova] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.302698] env[61857]: DEBUG nova.network.neutron [-] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.350830] env[61857]: DEBUG nova.network.neutron [req-eb36dd9c-5b74-4668-8fd2-f0311f69ca73 req-e894dd0d-adb6-4fad-ab7e-28f027b3c843 service nova] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.449457] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.449686] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.645849] env[61857]: DEBUG nova.compute.utils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.648893] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 687.649074] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 687.691858] env[61857]: DEBUG nova.policy [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a3a9944299d4574bf46978852f11d92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b21cc6a46cc04f20b942603dc80ec6c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.804742] env[61857]: INFO nova.compute.manager [-] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Took 1.03 seconds to deallocate network for instance. [ 687.811128] env[61857]: DEBUG nova.compute.claims [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 687.811128] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.829875] env[61857]: INFO nova.scheduler.client.report [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Deleted allocations for instance 95fbef26-78a7-4449-ba06-ffd517fb1032 [ 687.854526] env[61857]: DEBUG oslo_concurrency.lockutils [req-eb36dd9c-5b74-4668-8fd2-f0311f69ca73 req-e894dd0d-adb6-4fad-ab7e-28f027b3c843 service nova] Releasing lock "refresh_cache-51c54fe3-c0e2-4151-9626-cfb10cd7996e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.875364] env[61857]: DEBUG nova.compute.manager [req-abcaae32-eba3-4c53-8edd-ccb732e54349 req-d939981c-f159-4d0a-a529-ec43ccbbd172 service nova] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Received event network-vif-deleted-fe95c043-c85d-470c-ac22-755174721da1 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 687.998955] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Successfully created port: 9e56e450-a058-4ede-b734-ca42739a4736 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 688.152673] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.338470] env[61857]: DEBUG oslo_concurrency.lockutils [None req-66b82947-6ffc-43f3-b81e-ada4ac7da2fb tempest-ImagesOneServerNegativeTestJSON-959885766 tempest-ImagesOneServerNegativeTestJSON-959885766-project-member] Lock "95fbef26-78a7-4449-ba06-ffd517fb1032" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.172s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.572523] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db460da3-12f4-48e2-b8b3-b3421a2fdaf2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.580646] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45431b7c-6b72-42b2-b628-377f10158610 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.614505] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f08fa5-e3d9-4eec-8c61-39af6b51faf9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.622653] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2088bbe-1f1a-4839-b761-06f76ee49bf8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.639346] env[61857]: DEBUG nova.compute.provider_tree [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.841474] env[61857]: DEBUG nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 689.144124] env[61857]: DEBUG nova.scheduler.client.report [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.168486] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 689.192631] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.192877] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.193170] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.193512] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.194069] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.194612] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.194899] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.195133] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.195377] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.195599] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.195828] env[61857]: DEBUG nova.virt.hardware [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.197238] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b42e17-7c5d-45c4-ae32-cd1504868438 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.206448] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d455a279-9d31-469a-b356-d0ecdd669516 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.223150] env[61857]: ERROR nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. [ 689.223150] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.223150] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.223150] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.223150] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.223150] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.223150] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.223150] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.223150] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.223150] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 689.223150] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.223150] env[61857]: ERROR nova.compute.manager raise self.value [ 689.223150] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.223150] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.223150] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.223150] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.223644] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.223644] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.223644] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. [ 689.223644] env[61857]: ERROR nova.compute.manager [ 689.223644] env[61857]: Traceback (most recent call last): [ 689.223644] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.223644] env[61857]: listener.cb(fileno) [ 689.223644] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.223644] env[61857]: result = function(*args, **kwargs) [ 689.223644] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.223644] env[61857]: return func(*args, **kwargs) [ 689.223644] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.223644] env[61857]: raise e [ 689.223644] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.223644] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 689.223644] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.223644] env[61857]: created_port_ids = self._update_ports_for_instance( [ 689.223644] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.223644] env[61857]: with excutils.save_and_reraise_exception(): [ 689.223644] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.223644] env[61857]: self.force_reraise() [ 689.223644] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.223644] env[61857]: raise self.value [ 689.223644] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.223644] env[61857]: updated_port = self._update_port( [ 689.223644] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.223644] env[61857]: _ensure_no_port_binding_failure(port) [ 689.223644] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.223644] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.224570] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. [ 689.224570] env[61857]: Removing descriptor: 17 [ 689.224570] env[61857]: ERROR nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. [ 689.224570] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Traceback (most recent call last): [ 689.224570] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 689.224570] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] yield resources [ 689.224570] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.224570] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] self.driver.spawn(context, instance, image_meta, [ 689.224570] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.224570] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.224570] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.224570] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] vm_ref = self.build_virtual_machine(instance, [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] for vif in network_info: [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] return self._sync_wrapper(fn, *args, **kwargs) [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] self.wait() [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] self[:] = self._gt.wait() [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] return self._exit_event.wait() [ 689.224891] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] result = hub.switch() [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] return self.greenlet.switch() [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] result = function(*args, **kwargs) [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] return func(*args, **kwargs) [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] raise e [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] nwinfo = self.network_api.allocate_for_instance( [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.225322] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] created_port_ids = self._update_ports_for_instance( [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] with excutils.save_and_reraise_exception(): [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] self.force_reraise() [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] raise self.value [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] updated_port = self._update_port( [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] _ensure_no_port_binding_failure(port) [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.225714] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] raise exception.PortBindingFailed(port_id=port['id']) [ 689.226031] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] nova.exception.PortBindingFailed: Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. [ 689.226031] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] [ 689.226031] env[61857]: INFO nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Terminating instance [ 689.227441] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "refresh_cache-8ee206cf-c906-4ac0-8ba7-d864f5f538db" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.227604] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired lock "refresh_cache-8ee206cf-c906-4ac0-8ba7-d864f5f538db" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.227773] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 689.363135] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.647011] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.648145] env[61857]: DEBUG nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 689.656021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.640s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.656021] env[61857]: INFO nova.compute.claims [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.767406] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.921356] env[61857]: DEBUG nova.compute.manager [req-73045c88-f16c-41af-be95-d1fff9233fa1 req-779a9bd4-0134-4aee-a95c-a18d018ef9ba service nova] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Received event network-changed-9e56e450-a058-4ede-b734-ca42739a4736 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 689.921412] env[61857]: DEBUG nova.compute.manager [req-73045c88-f16c-41af-be95-d1fff9233fa1 req-779a9bd4-0134-4aee-a95c-a18d018ef9ba service nova] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Refreshing instance network info cache due to event network-changed-9e56e450-a058-4ede-b734-ca42739a4736. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 689.921561] env[61857]: DEBUG oslo_concurrency.lockutils [req-73045c88-f16c-41af-be95-d1fff9233fa1 req-779a9bd4-0134-4aee-a95c-a18d018ef9ba service nova] Acquiring lock "refresh_cache-8ee206cf-c906-4ac0-8ba7-d864f5f538db" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.922696] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.156731] env[61857]: DEBUG nova.compute.utils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.159935] env[61857]: DEBUG nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 690.160088] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 690.250921] env[61857]: DEBUG nova.policy [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1496582f8474b14baf63b215a80e64e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3688a3ca85044527a84ad1a7b201fbdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 690.424722] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Releasing lock "refresh_cache-8ee206cf-c906-4ac0-8ba7-d864f5f538db" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.425212] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 690.425415] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 690.425743] env[61857]: DEBUG oslo_concurrency.lockutils [req-73045c88-f16c-41af-be95-d1fff9233fa1 req-779a9bd4-0134-4aee-a95c-a18d018ef9ba service nova] Acquired lock "refresh_cache-8ee206cf-c906-4ac0-8ba7-d864f5f538db" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.425911] env[61857]: DEBUG nova.network.neutron [req-73045c88-f16c-41af-be95-d1fff9233fa1 req-779a9bd4-0134-4aee-a95c-a18d018ef9ba service nova] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Refreshing network info cache for port 9e56e450-a058-4ede-b734-ca42739a4736 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 690.426993] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5c40e998-b2d8-4392-a20f-becc29a01c41 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.441349] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcb0979-7437-4c8b-a61d-0dc5061d386a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.474605] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8ee206cf-c906-4ac0-8ba7-d864f5f538db could not be found. [ 690.475110] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 690.475372] env[61857]: INFO nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Took 0.05 seconds to destroy the instance on the hypervisor. [ 690.475914] env[61857]: DEBUG oslo.service.loopingcall [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.476197] env[61857]: DEBUG nova.compute.manager [-] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.476299] env[61857]: DEBUG nova.network.neutron [-] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 690.493414] env[61857]: DEBUG nova.network.neutron [-] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.619660] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Successfully created port: 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.661510] env[61857]: DEBUG nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 690.952380] env[61857]: DEBUG nova.network.neutron [req-73045c88-f16c-41af-be95-d1fff9233fa1 req-779a9bd4-0134-4aee-a95c-a18d018ef9ba service nova] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.996274] env[61857]: DEBUG nova.network.neutron [-] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.073701] env[61857]: DEBUG nova.network.neutron [req-73045c88-f16c-41af-be95-d1fff9233fa1 req-779a9bd4-0134-4aee-a95c-a18d018ef9ba service nova] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.149879] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b996d568-5cc8-4e1b-a6fe-387dde3b5b35 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.159273] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d078768-f83a-4231-917e-2b0c56805173 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.193591] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406e4ceb-7fa1-4a45-9bb5-39f3f94d76e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.202652] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b940ce16-9ef3-49f5-8ba0-cc057df74072 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.219363] env[61857]: DEBUG nova.compute.provider_tree [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.486944] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.487202] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.498699] env[61857]: INFO nova.compute.manager [-] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Took 1.02 seconds to deallocate network for instance. [ 691.504331] env[61857]: DEBUG nova.compute.claims [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 691.504487] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.576613] env[61857]: DEBUG oslo_concurrency.lockutils [req-73045c88-f16c-41af-be95-d1fff9233fa1 req-779a9bd4-0134-4aee-a95c-a18d018ef9ba service nova] Releasing lock "refresh_cache-8ee206cf-c906-4ac0-8ba7-d864f5f538db" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.576900] env[61857]: DEBUG nova.compute.manager [req-73045c88-f16c-41af-be95-d1fff9233fa1 req-779a9bd4-0134-4aee-a95c-a18d018ef9ba service nova] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Received event network-vif-deleted-9e56e450-a058-4ede-b734-ca42739a4736 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 691.694462] env[61857]: DEBUG nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 691.724019] env[61857]: DEBUG nova.scheduler.client.report [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 691.729554] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 691.729775] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 691.730034] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.730144] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 691.730312] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.730462] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 691.730663] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 691.730821] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 691.731025] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 691.731187] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 691.731395] env[61857]: DEBUG nova.virt.hardware [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.732262] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab45092-8787-4703-989b-b6c5d07540f7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.743582] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c05baf2-bde0-4825-a9ab-582a2b7ca84a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.952426] env[61857]: DEBUG nova.compute.manager [req-fbd54972-73a4-4a6e-98b4-3d24eea6ce06 req-d14105ac-3741-46c3-9c60-4c23c9c1baed service nova] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Received event network-changed-967189fc-2ecf-4f90-ab01-8af7ce3fe2c6 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 691.952426] env[61857]: DEBUG nova.compute.manager [req-fbd54972-73a4-4a6e-98b4-3d24eea6ce06 req-d14105ac-3741-46c3-9c60-4c23c9c1baed service nova] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Refreshing instance network info cache due to event network-changed-967189fc-2ecf-4f90-ab01-8af7ce3fe2c6. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 691.952426] env[61857]: DEBUG oslo_concurrency.lockutils [req-fbd54972-73a4-4a6e-98b4-3d24eea6ce06 req-d14105ac-3741-46c3-9c60-4c23c9c1baed service nova] Acquiring lock "refresh_cache-ca3d3ad9-39a0-467b-812f-21b7303b8bc8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.952426] env[61857]: DEBUG oslo_concurrency.lockutils [req-fbd54972-73a4-4a6e-98b4-3d24eea6ce06 req-d14105ac-3741-46c3-9c60-4c23c9c1baed service nova] Acquired lock "refresh_cache-ca3d3ad9-39a0-467b-812f-21b7303b8bc8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.952734] env[61857]: DEBUG nova.network.neutron [req-fbd54972-73a4-4a6e-98b4-3d24eea6ce06 req-d14105ac-3741-46c3-9c60-4c23c9c1baed service nova] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Refreshing network info cache for port 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 692.026412] env[61857]: ERROR nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6, please check neutron logs for more information. [ 692.026412] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 692.026412] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.026412] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 692.026412] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.026412] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 692.026412] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.026412] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 692.026412] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.026412] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 692.026412] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.026412] env[61857]: ERROR nova.compute.manager raise self.value [ 692.026412] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.026412] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 692.026412] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.026412] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 692.026936] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.026936] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 692.026936] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6, please check neutron logs for more information. [ 692.026936] env[61857]: ERROR nova.compute.manager [ 692.026936] env[61857]: Traceback (most recent call last): [ 692.026936] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 692.026936] env[61857]: listener.cb(fileno) [ 692.026936] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.026936] env[61857]: result = function(*args, **kwargs) [ 692.026936] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.026936] env[61857]: return func(*args, **kwargs) [ 692.026936] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.026936] env[61857]: raise e [ 692.026936] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.026936] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 692.026936] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.026936] env[61857]: created_port_ids = self._update_ports_for_instance( [ 692.026936] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.026936] env[61857]: with excutils.save_and_reraise_exception(): [ 692.026936] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.026936] env[61857]: self.force_reraise() [ 692.026936] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.026936] env[61857]: raise self.value [ 692.026936] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.026936] env[61857]: updated_port = self._update_port( [ 692.026936] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.026936] env[61857]: _ensure_no_port_binding_failure(port) [ 692.026936] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.026936] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 692.027799] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6, please check neutron logs for more information. [ 692.027799] env[61857]: Removing descriptor: 17 [ 692.027799] env[61857]: ERROR nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6, please check neutron logs for more information. [ 692.027799] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Traceback (most recent call last): [ 692.027799] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 692.027799] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] yield resources [ 692.027799] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.027799] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] self.driver.spawn(context, instance, image_meta, [ 692.027799] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 692.027799] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.027799] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.027799] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] vm_ref = self.build_virtual_machine(instance, [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] for vif in network_info: [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] return self._sync_wrapper(fn, *args, **kwargs) [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] self.wait() [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] self[:] = self._gt.wait() [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] return self._exit_event.wait() [ 692.028187] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] result = hub.switch() [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] return self.greenlet.switch() [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] result = function(*args, **kwargs) [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] return func(*args, **kwargs) [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] raise e [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] nwinfo = self.network_api.allocate_for_instance( [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.028964] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] created_port_ids = self._update_ports_for_instance( [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] with excutils.save_and_reraise_exception(): [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] self.force_reraise() [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] raise self.value [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] updated_port = self._update_port( [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] _ensure_no_port_binding_failure(port) [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.029411] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] raise exception.PortBindingFailed(port_id=port['id']) [ 692.029745] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] nova.exception.PortBindingFailed: Binding failed for port 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6, please check neutron logs for more information. [ 692.029745] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] [ 692.029745] env[61857]: INFO nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Terminating instance [ 692.029917] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "refresh_cache-ca3d3ad9-39a0-467b-812f-21b7303b8bc8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.041292] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.041519] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.236614] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.237161] env[61857]: DEBUG nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.240435] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.156s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.241989] env[61857]: INFO nova.compute.claims [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.471437] env[61857]: DEBUG nova.network.neutron [req-fbd54972-73a4-4a6e-98b4-3d24eea6ce06 req-d14105ac-3741-46c3-9c60-4c23c9c1baed service nova] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.574991] env[61857]: DEBUG nova.network.neutron [req-fbd54972-73a4-4a6e-98b4-3d24eea6ce06 req-d14105ac-3741-46c3-9c60-4c23c9c1baed service nova] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.747057] env[61857]: DEBUG nova.compute.utils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 692.750393] env[61857]: DEBUG nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 692.750628] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 692.789678] env[61857]: DEBUG nova.policy [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c23e95cd2584bada69e98ba9e8575ab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '180bf8932fb845c68327ab92ead371d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.081123] env[61857]: DEBUG oslo_concurrency.lockutils [req-fbd54972-73a4-4a6e-98b4-3d24eea6ce06 req-d14105ac-3741-46c3-9c60-4c23c9c1baed service nova] Releasing lock "refresh_cache-ca3d3ad9-39a0-467b-812f-21b7303b8bc8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.081664] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "refresh_cache-ca3d3ad9-39a0-467b-812f-21b7303b8bc8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.081860] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 693.122128] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Successfully created port: 815a603a-b57e-4bca-823f-635e83967652 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.256021] env[61857]: DEBUG nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 693.620233] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.737013] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.775946] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a16074-f5b3-47f3-9dee-1a634d6b1d7f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.787100] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9416136-b626-45ca-917c-d90f10b17cc9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.822584] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9876b227-c45a-478b-abcd-a1f1561f4f5a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.831409] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c24c022-8118-44e2-86a5-737159c0bb3a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.848771] env[61857]: DEBUG nova.compute.provider_tree [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.076633] env[61857]: DEBUG nova.compute.manager [req-b4d0f4bb-2845-478c-8acf-d682e772f74e req-43904ae8-a898-4d19-ae96-c0582bcd382f service nova] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Received event network-vif-deleted-967189fc-2ecf-4f90-ab01-8af7ce3fe2c6 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 694.245141] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "refresh_cache-ca3d3ad9-39a0-467b-812f-21b7303b8bc8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.245624] env[61857]: DEBUG nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.245820] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 694.246151] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b3dd7b74-8a2f-484b-8440-532b037aa055 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.263536] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34c4504-b4fd-414e-aebc-eb46a4398f37 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.276410] env[61857]: DEBUG nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.289478] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ca3d3ad9-39a0-467b-812f-21b7303b8bc8 could not be found. [ 694.289704] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 694.289881] env[61857]: INFO nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 694.290178] env[61857]: DEBUG oslo.service.loopingcall [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.290462] env[61857]: DEBUG nova.compute.manager [-] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.290562] env[61857]: DEBUG nova.network.neutron [-] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 694.317024] env[61857]: DEBUG nova.network.neutron [-] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 694.320335] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.320738] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.321580] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.321833] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.322067] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.322404] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.322934] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.323154] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.323340] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.323534] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.323739] env[61857]: DEBUG nova.virt.hardware [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.324608] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab41c87a-dd5e-4463-b21b-3ae013cc81ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.333506] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d7cf0b-1db2-42fa-b57a-52c9b2ef83b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.352040] env[61857]: DEBUG nova.scheduler.client.report [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.428235] env[61857]: ERROR nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 815a603a-b57e-4bca-823f-635e83967652, please check neutron logs for more information. [ 694.428235] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.428235] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.428235] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.428235] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.428235] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.428235] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.428235] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.428235] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.428235] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 694.428235] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.428235] env[61857]: ERROR nova.compute.manager raise self.value [ 694.428235] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.428235] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.428235] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.428235] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.428803] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.428803] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.428803] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 815a603a-b57e-4bca-823f-635e83967652, please check neutron logs for more information. [ 694.428803] env[61857]: ERROR nova.compute.manager [ 694.428803] env[61857]: Traceback (most recent call last): [ 694.428803] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.428803] env[61857]: listener.cb(fileno) [ 694.428803] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.428803] env[61857]: result = function(*args, **kwargs) [ 694.428803] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.428803] env[61857]: return func(*args, **kwargs) [ 694.428803] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.428803] env[61857]: raise e [ 694.428803] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.428803] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 694.428803] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.428803] env[61857]: created_port_ids = self._update_ports_for_instance( [ 694.428803] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.428803] env[61857]: with excutils.save_and_reraise_exception(): [ 694.428803] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.428803] env[61857]: self.force_reraise() [ 694.428803] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.428803] env[61857]: raise self.value [ 694.428803] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.428803] env[61857]: updated_port = self._update_port( [ 694.428803] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.428803] env[61857]: _ensure_no_port_binding_failure(port) [ 694.428803] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.428803] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.429622] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 815a603a-b57e-4bca-823f-635e83967652, please check neutron logs for more information. [ 694.429622] env[61857]: Removing descriptor: 17 [ 694.429622] env[61857]: ERROR nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 815a603a-b57e-4bca-823f-635e83967652, please check neutron logs for more information. [ 694.429622] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Traceback (most recent call last): [ 694.429622] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 694.429622] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] yield resources [ 694.429622] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.429622] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] self.driver.spawn(context, instance, image_meta, [ 694.429622] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 694.429622] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.429622] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.429622] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] vm_ref = self.build_virtual_machine(instance, [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] for vif in network_info: [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] return self._sync_wrapper(fn, *args, **kwargs) [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] self.wait() [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] self[:] = self._gt.wait() [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] return self._exit_event.wait() [ 694.430253] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] result = hub.switch() [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] return self.greenlet.switch() [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] result = function(*args, **kwargs) [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] return func(*args, **kwargs) [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] raise e [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] nwinfo = self.network_api.allocate_for_instance( [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.430591] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] created_port_ids = self._update_ports_for_instance( [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] with excutils.save_and_reraise_exception(): [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] self.force_reraise() [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] raise self.value [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] updated_port = self._update_port( [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] _ensure_no_port_binding_failure(port) [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.430935] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] raise exception.PortBindingFailed(port_id=port['id']) [ 694.431406] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] nova.exception.PortBindingFailed: Binding failed for port 815a603a-b57e-4bca-823f-635e83967652, please check neutron logs for more information. [ 694.431406] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] [ 694.431406] env[61857]: INFO nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Terminating instance [ 694.433929] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "refresh_cache-9edf9e8c-5421-4018-afee-59e10cd335ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.433929] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquired lock "refresh_cache-9edf9e8c-5421-4018-afee-59e10cd335ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.433929] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 694.830928] env[61857]: DEBUG nova.network.neutron [-] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.858071] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.858475] env[61857]: DEBUG nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 694.865348] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.876s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.951682] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.055523] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.337336] env[61857]: INFO nova.compute.manager [-] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Took 1.05 seconds to deallocate network for instance. [ 695.342103] env[61857]: DEBUG nova.compute.claims [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 695.342647] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.374023] env[61857]: DEBUG nova.compute.utils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.378026] env[61857]: DEBUG nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 695.378240] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 695.428609] env[61857]: DEBUG nova.policy [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18e11b5eaf09476caeab46742f3b7859', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e019f307dcd406d8c59edfc1596cc86', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 695.561627] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Releasing lock "refresh_cache-9edf9e8c-5421-4018-afee-59e10cd335ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.562589] env[61857]: DEBUG nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 695.563023] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 695.563782] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8bc94579-b862-48ae-855f-d553e1afe6e1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.576181] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddd55e2-99e2-42bc-8e74-bba3f67fd6bd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.607185] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9edf9e8c-5421-4018-afee-59e10cd335ef could not be found. [ 695.607413] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 695.607591] env[61857]: INFO nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Took 0.04 seconds to destroy the instance on the hypervisor. [ 695.607832] env[61857]: DEBUG oslo.service.loopingcall [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.611271] env[61857]: DEBUG nova.compute.manager [-] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.611377] env[61857]: DEBUG nova.network.neutron [-] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 695.628684] env[61857]: DEBUG nova.network.neutron [-] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.801367] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Successfully created port: 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 695.816055] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f015af8-6cda-4eb7-a895-fb6e2bdcc2df {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.823857] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a156722-f817-4c15-96c0-4448ae27dfff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.864938] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8143182d-3b91-420d-94a8-113ee3e37410 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.876350] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec96ef02-9eb0-4ff4-824d-b6dc1f0b225d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.881203] env[61857]: DEBUG nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 695.894417] env[61857]: DEBUG nova.compute.provider_tree [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.978494] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.978737] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.133190] env[61857]: DEBUG nova.network.neutron [-] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.143036] env[61857]: DEBUG nova.compute.manager [req-05db79bf-1569-4025-8dc2-1b5c9afecb50 req-cafd31ee-6a97-4cf9-9c08-3d00dc9f40d8 service nova] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Received event network-changed-815a603a-b57e-4bca-823f-635e83967652 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 696.143036] env[61857]: DEBUG nova.compute.manager [req-05db79bf-1569-4025-8dc2-1b5c9afecb50 req-cafd31ee-6a97-4cf9-9c08-3d00dc9f40d8 service nova] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Refreshing instance network info cache due to event network-changed-815a603a-b57e-4bca-823f-635e83967652. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 696.143036] env[61857]: DEBUG oslo_concurrency.lockutils [req-05db79bf-1569-4025-8dc2-1b5c9afecb50 req-cafd31ee-6a97-4cf9-9c08-3d00dc9f40d8 service nova] Acquiring lock "refresh_cache-9edf9e8c-5421-4018-afee-59e10cd335ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.143036] env[61857]: DEBUG oslo_concurrency.lockutils [req-05db79bf-1569-4025-8dc2-1b5c9afecb50 req-cafd31ee-6a97-4cf9-9c08-3d00dc9f40d8 service nova] Acquired lock "refresh_cache-9edf9e8c-5421-4018-afee-59e10cd335ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.143291] env[61857]: DEBUG nova.network.neutron [req-05db79bf-1569-4025-8dc2-1b5c9afecb50 req-cafd31ee-6a97-4cf9-9c08-3d00dc9f40d8 service nova] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Refreshing network info cache for port 815a603a-b57e-4bca-823f-635e83967652 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 696.399850] env[61857]: INFO nova.virt.block_device [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Booting with volume 3c2bdde8-7292-46ea-b99b-ba0a410f17a8 at /dev/sda [ 696.401054] env[61857]: DEBUG nova.scheduler.client.report [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.464019] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-afc0d558-9174-4cc7-82fd-82312bf29480 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.475020] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ba3d4a-9aec-44f2-b142-3208b3821836 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.504263] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-251eae23-41b4-42e7-8f8b-ee819cbaeeb5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.511613] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad925cc-ebf1-47c5-a2c7-c2400e7a5941 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.532615] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e36de4eb-b937-4cbe-ad52-a2d51e5863fe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.538589] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93eb888-3c2d-4e31-b51b-5c12a653f8b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.556377] env[61857]: DEBUG nova.virt.block_device [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Updating existing volume attachment record: 64c48bf6-03c1-4308-b3b3-9deaaf3143f6 {{(pid=61857) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 696.635070] env[61857]: INFO nova.compute.manager [-] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Took 1.02 seconds to deallocate network for instance. [ 696.637488] env[61857]: DEBUG nova.compute.claims [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 696.637673] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.665825] env[61857]: DEBUG nova.network.neutron [req-05db79bf-1569-4025-8dc2-1b5c9afecb50 req-cafd31ee-6a97-4cf9-9c08-3d00dc9f40d8 service nova] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.768536] env[61857]: DEBUG nova.network.neutron [req-05db79bf-1569-4025-8dc2-1b5c9afecb50 req-cafd31ee-6a97-4cf9-9c08-3d00dc9f40d8 service nova] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.906610] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.045s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.907312] env[61857]: ERROR nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9069d1ca-2999-4d67-a1ca-eaa5b68f051b, please check neutron logs for more information. [ 696.907312] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Traceback (most recent call last): [ 696.907312] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.907312] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] self.driver.spawn(context, instance, image_meta, [ 696.907312] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 696.907312] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.907312] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.907312] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] vm_ref = self.build_virtual_machine(instance, [ 696.907312] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.907312] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.907312] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] for vif in network_info: [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] return self._sync_wrapper(fn, *args, **kwargs) [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] self.wait() [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] self[:] = self._gt.wait() [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] return self._exit_event.wait() [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] result = hub.switch() [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 696.907675] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] return self.greenlet.switch() [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] result = function(*args, **kwargs) [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] return func(*args, **kwargs) [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] raise e [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] nwinfo = self.network_api.allocate_for_instance( [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] created_port_ids = self._update_ports_for_instance( [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] with excutils.save_and_reraise_exception(): [ 696.907997] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] self.force_reraise() [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] raise self.value [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] updated_port = self._update_port( [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] _ensure_no_port_binding_failure(port) [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] raise exception.PortBindingFailed(port_id=port['id']) [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] nova.exception.PortBindingFailed: Binding failed for port 9069d1ca-2999-4d67-a1ca-eaa5b68f051b, please check neutron logs for more information. [ 696.908447] env[61857]: ERROR nova.compute.manager [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] [ 696.908771] env[61857]: DEBUG nova.compute.utils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Binding failed for port 9069d1ca-2999-4d67-a1ca-eaa5b68f051b, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 696.909365] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.277s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.911598] env[61857]: INFO nova.compute.claims [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.914418] env[61857]: DEBUG nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Build of instance 4ede4082-93b0-4cac-9e28-5448f04cba9b was re-scheduled: Binding failed for port 9069d1ca-2999-4d67-a1ca-eaa5b68f051b, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 696.914915] env[61857]: DEBUG nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 696.915202] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Acquiring lock "refresh_cache-4ede4082-93b0-4cac-9e28-5448f04cba9b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.915496] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Acquired lock "refresh_cache-4ede4082-93b0-4cac-9e28-5448f04cba9b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.915565] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.087770] env[61857]: ERROR nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b, please check neutron logs for more information. [ 697.087770] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 697.087770] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.087770] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 697.087770] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.087770] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 697.087770] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.087770] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 697.087770] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.087770] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 697.087770] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.087770] env[61857]: ERROR nova.compute.manager raise self.value [ 697.087770] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.087770] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 697.087770] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.087770] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 697.088264] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.088264] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 697.088264] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b, please check neutron logs for more information. [ 697.088264] env[61857]: ERROR nova.compute.manager [ 697.088264] env[61857]: Traceback (most recent call last): [ 697.088264] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 697.088264] env[61857]: listener.cb(fileno) [ 697.088264] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.088264] env[61857]: result = function(*args, **kwargs) [ 697.088264] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.088264] env[61857]: return func(*args, **kwargs) [ 697.088264] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.088264] env[61857]: raise e [ 697.088264] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.088264] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 697.088264] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.088264] env[61857]: created_port_ids = self._update_ports_for_instance( [ 697.088264] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.088264] env[61857]: with excutils.save_and_reraise_exception(): [ 697.088264] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.088264] env[61857]: self.force_reraise() [ 697.088264] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.088264] env[61857]: raise self.value [ 697.088264] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.088264] env[61857]: updated_port = self._update_port( [ 697.088264] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.088264] env[61857]: _ensure_no_port_binding_failure(port) [ 697.088264] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.088264] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 697.089380] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b, please check neutron logs for more information. [ 697.089380] env[61857]: Removing descriptor: 17 [ 697.270807] env[61857]: DEBUG oslo_concurrency.lockutils [req-05db79bf-1569-4025-8dc2-1b5c9afecb50 req-cafd31ee-6a97-4cf9-9c08-3d00dc9f40d8 service nova] Releasing lock "refresh_cache-9edf9e8c-5421-4018-afee-59e10cd335ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.271099] env[61857]: DEBUG nova.compute.manager [req-05db79bf-1569-4025-8dc2-1b5c9afecb50 req-cafd31ee-6a97-4cf9-9c08-3d00dc9f40d8 service nova] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Received event network-vif-deleted-815a603a-b57e-4bca-823f-635e83967652 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 697.434877] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.527694] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.029874] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Releasing lock "refresh_cache-4ede4082-93b0-4cac-9e28-5448f04cba9b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.030148] env[61857]: DEBUG nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 698.030343] env[61857]: DEBUG nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.030563] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 698.046338] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.173017] env[61857]: DEBUG nova.compute.manager [req-3530aee8-f1ba-469e-95e2-005ca8666bc9 req-5c81ad91-42fc-45de-a876-091136a8fd70 service nova] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Received event network-changed-0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 698.173017] env[61857]: DEBUG nova.compute.manager [req-3530aee8-f1ba-469e-95e2-005ca8666bc9 req-5c81ad91-42fc-45de-a876-091136a8fd70 service nova] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Refreshing instance network info cache due to event network-changed-0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 698.173017] env[61857]: DEBUG oslo_concurrency.lockutils [req-3530aee8-f1ba-469e-95e2-005ca8666bc9 req-5c81ad91-42fc-45de-a876-091136a8fd70 service nova] Acquiring lock "refresh_cache-e52972ca-877f-45b8-b79f-c834246a8857" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.173017] env[61857]: DEBUG oslo_concurrency.lockutils [req-3530aee8-f1ba-469e-95e2-005ca8666bc9 req-5c81ad91-42fc-45de-a876-091136a8fd70 service nova] Acquired lock "refresh_cache-e52972ca-877f-45b8-b79f-c834246a8857" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.173017] env[61857]: DEBUG nova.network.neutron [req-3530aee8-f1ba-469e-95e2-005ca8666bc9 req-5c81ad91-42fc-45de-a876-091136a8fd70 service nova] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Refreshing network info cache for port 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 698.275507] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaee736b-2118-4d49-bfaf-8e1ec0b7e9b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.283560] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5fb692e-294c-41b1-8c9e-707d41cb17fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.314571] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a714842-dfc5-4dc1-9057-8835b334a201 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.321661] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a5f182-a152-4141-90f4-812eb106a802 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.334515] env[61857]: DEBUG nova.compute.provider_tree [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.548801] env[61857]: DEBUG nova.network.neutron [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.690354] env[61857]: DEBUG nova.network.neutron [req-3530aee8-f1ba-469e-95e2-005ca8666bc9 req-5c81ad91-42fc-45de-a876-091136a8fd70 service nova] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.693105] env[61857]: DEBUG nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 698.693687] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 698.693953] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 698.694175] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 698.694422] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 698.694620] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 698.694814] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 698.695095] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 698.695316] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 698.695533] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 698.695745] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 698.695970] env[61857]: DEBUG nova.virt.hardware [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 698.696927] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e6fcb3-212b-4742-927f-e213eedcf2e2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.706898] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7c6d82-e0fc-4268-98f4-48ddae709619 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.722835] env[61857]: ERROR nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b, please check neutron logs for more information. [ 698.722835] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] Traceback (most recent call last): [ 698.722835] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 698.722835] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] yield resources [ 698.722835] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 698.722835] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] self.driver.spawn(context, instance, image_meta, [ 698.722835] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 698.722835] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] self._vmops.spawn(context, instance, image_meta, injected_files, [ 698.722835] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 698.722835] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] vm_ref = self.build_virtual_machine(instance, [ 698.722835] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] vif_infos = vmwarevif.get_vif_info(self._session, [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] for vif in network_info: [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] return self._sync_wrapper(fn, *args, **kwargs) [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] self.wait() [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] self[:] = self._gt.wait() [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] return self._exit_event.wait() [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 698.723300] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] current.throw(*self._exc) [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] result = function(*args, **kwargs) [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] return func(*args, **kwargs) [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] raise e [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] nwinfo = self.network_api.allocate_for_instance( [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] created_port_ids = self._update_ports_for_instance( [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] with excutils.save_and_reraise_exception(): [ 698.723771] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] self.force_reraise() [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] raise self.value [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] updated_port = self._update_port( [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] _ensure_no_port_binding_failure(port) [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] raise exception.PortBindingFailed(port_id=port['id']) [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] nova.exception.PortBindingFailed: Binding failed for port 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b, please check neutron logs for more information. [ 698.727060] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] [ 698.727060] env[61857]: INFO nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Terminating instance [ 698.727478] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Acquiring lock "refresh_cache-e52972ca-877f-45b8-b79f-c834246a8857" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.799061] env[61857]: DEBUG nova.network.neutron [req-3530aee8-f1ba-469e-95e2-005ca8666bc9 req-5c81ad91-42fc-45de-a876-091136a8fd70 service nova] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.838028] env[61857]: DEBUG nova.scheduler.client.report [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.052051] env[61857]: INFO nova.compute.manager [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] [instance: 4ede4082-93b0-4cac-9e28-5448f04cba9b] Took 1.02 seconds to deallocate network for instance. [ 699.301744] env[61857]: DEBUG oslo_concurrency.lockutils [req-3530aee8-f1ba-469e-95e2-005ca8666bc9 req-5c81ad91-42fc-45de-a876-091136a8fd70 service nova] Releasing lock "refresh_cache-e52972ca-877f-45b8-b79f-c834246a8857" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.302039] env[61857]: DEBUG nova.compute.manager [req-3530aee8-f1ba-469e-95e2-005ca8666bc9 req-5c81ad91-42fc-45de-a876-091136a8fd70 service nova] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Received event network-vif-deleted-0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 699.302401] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Acquired lock "refresh_cache-e52972ca-877f-45b8-b79f-c834246a8857" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.302645] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 699.343248] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.343783] env[61857]: DEBUG nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 699.347024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.690s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.347962] env[61857]: INFO nova.compute.claims [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 699.824862] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.848408] env[61857]: DEBUG nova.compute.utils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.849790] env[61857]: DEBUG nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 699.849958] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 699.906761] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.909242] env[61857]: DEBUG nova.policy [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd28e2a95200a41c1bd11d47e8d8fb6ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46d84780cba746b0814fedff9d18090c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 700.084966] env[61857]: INFO nova.scheduler.client.report [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Deleted allocations for instance 4ede4082-93b0-4cac-9e28-5448f04cba9b [ 700.197325] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Successfully created port: 51b2e7bb-94cc-4ae3-a62a-e860d82899f9 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.353507] env[61857]: DEBUG nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 700.412495] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Releasing lock "refresh_cache-e52972ca-877f-45b8-b79f-c834246a8857" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.413043] env[61857]: DEBUG nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 700.413358] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0865f5ba-0669-46d5-b1aa-5d478a933fdb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.425616] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0a2e26-76c1-4317-a08a-34cd8ea7fd10 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.451296] env[61857]: WARNING nova.virt.vmwareapi.driver [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance e52972ca-877f-45b8-b79f-c834246a8857 could not be found. [ 700.451296] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 700.451598] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-47c3a5bc-5a57-477b-b97e-5d0f45ec77c5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.458964] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c810134-1f97-4861-926a-b5110f1b4b1d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.491284] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e52972ca-877f-45b8-b79f-c834246a8857 could not be found. [ 700.491517] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 700.491696] env[61857]: INFO nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Took 0.08 seconds to destroy the instance on the hypervisor. [ 700.491944] env[61857]: DEBUG oslo.service.loopingcall [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.493054] env[61857]: DEBUG nova.compute.manager [-] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.493054] env[61857]: DEBUG nova.network.neutron [-] [instance: e52972ca-877f-45b8-b79f-c834246a8857] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 700.506208] env[61857]: DEBUG nova.network.neutron [-] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.596929] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a70cb02d-ef24-4aca-b672-15bb2ee6044c tempest-FloatingIPsAssociationNegativeTestJSON-1865368323 tempest-FloatingIPsAssociationNegativeTestJSON-1865368323-project-member] Lock "4ede4082-93b0-4cac-9e28-5448f04cba9b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.543s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.870445] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b25d2e9-3016-4b90-af5a-31141651c57b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.882616] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667bc203-a5e5-4a87-952b-1b51f94f9a35 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.922160] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a59ba8-c267-4128-860d-35b0c83ed6ee {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.930986] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3882301-c118-4d55-8cb7-c77df4cbbded {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.946709] env[61857]: DEBUG nova.compute.provider_tree [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.008671] env[61857]: DEBUG nova.network.neutron [-] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.112810] env[61857]: DEBUG nova.compute.manager [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.202327] env[61857]: DEBUG nova.compute.manager [req-e66a92f0-40fc-497f-ac33-8469d7b75dcc req-428fd910-b273-4533-a838-8afcfe0a8734 service nova] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Received event network-changed-51b2e7bb-94cc-4ae3-a62a-e860d82899f9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 701.202564] env[61857]: DEBUG nova.compute.manager [req-e66a92f0-40fc-497f-ac33-8469d7b75dcc req-428fd910-b273-4533-a838-8afcfe0a8734 service nova] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Refreshing instance network info cache due to event network-changed-51b2e7bb-94cc-4ae3-a62a-e860d82899f9. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 701.202714] env[61857]: DEBUG oslo_concurrency.lockutils [req-e66a92f0-40fc-497f-ac33-8469d7b75dcc req-428fd910-b273-4533-a838-8afcfe0a8734 service nova] Acquiring lock "refresh_cache-952d0a1c-3e94-42dc-9bc4-8063275afc82" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.202858] env[61857]: DEBUG oslo_concurrency.lockutils [req-e66a92f0-40fc-497f-ac33-8469d7b75dcc req-428fd910-b273-4533-a838-8afcfe0a8734 service nova] Acquired lock "refresh_cache-952d0a1c-3e94-42dc-9bc4-8063275afc82" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.203035] env[61857]: DEBUG nova.network.neutron [req-e66a92f0-40fc-497f-ac33-8469d7b75dcc req-428fd910-b273-4533-a838-8afcfe0a8734 service nova] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Refreshing network info cache for port 51b2e7bb-94cc-4ae3-a62a-e860d82899f9 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 701.244765] env[61857]: ERROR nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 51b2e7bb-94cc-4ae3-a62a-e860d82899f9, please check neutron logs for more information. [ 701.244765] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.244765] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.244765] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.244765] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.244765] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.244765] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.244765] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.244765] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.244765] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 701.244765] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.244765] env[61857]: ERROR nova.compute.manager raise self.value [ 701.244765] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.244765] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.244765] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.244765] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.245245] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.245245] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.245245] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 51b2e7bb-94cc-4ae3-a62a-e860d82899f9, please check neutron logs for more information. [ 701.245245] env[61857]: ERROR nova.compute.manager [ 701.245245] env[61857]: Traceback (most recent call last): [ 701.245245] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.245245] env[61857]: listener.cb(fileno) [ 701.245245] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.245245] env[61857]: result = function(*args, **kwargs) [ 701.245245] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.245245] env[61857]: return func(*args, **kwargs) [ 701.245245] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.245245] env[61857]: raise e [ 701.245245] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.245245] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 701.245245] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.245245] env[61857]: created_port_ids = self._update_ports_for_instance( [ 701.245245] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.245245] env[61857]: with excutils.save_and_reraise_exception(): [ 701.245245] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.245245] env[61857]: self.force_reraise() [ 701.245245] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.245245] env[61857]: raise self.value [ 701.245245] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.245245] env[61857]: updated_port = self._update_port( [ 701.245245] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.245245] env[61857]: _ensure_no_port_binding_failure(port) [ 701.245245] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.245245] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.246171] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 51b2e7bb-94cc-4ae3-a62a-e860d82899f9, please check neutron logs for more information. [ 701.246171] env[61857]: Removing descriptor: 16 [ 701.375447] env[61857]: DEBUG nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 701.403764] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.403993] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.404197] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.404477] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.406072] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.406072] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.406072] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.406072] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.406072] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.406417] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.406417] env[61857]: DEBUG nova.virt.hardware [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.406417] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91dc8dd-068c-410e-9b8c-e19bac308e63 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.414398] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545a9db6-cb4e-4cb7-906f-1aa507795801 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.430454] env[61857]: ERROR nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 51b2e7bb-94cc-4ae3-a62a-e860d82899f9, please check neutron logs for more information. [ 701.430454] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Traceback (most recent call last): [ 701.430454] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 701.430454] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] yield resources [ 701.430454] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.430454] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] self.driver.spawn(context, instance, image_meta, [ 701.430454] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 701.430454] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.430454] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.430454] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] vm_ref = self.build_virtual_machine(instance, [ 701.430454] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] for vif in network_info: [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] return self._sync_wrapper(fn, *args, **kwargs) [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] self.wait() [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] self[:] = self._gt.wait() [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] return self._exit_event.wait() [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.430972] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] current.throw(*self._exc) [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] result = function(*args, **kwargs) [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] return func(*args, **kwargs) [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] raise e [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] nwinfo = self.network_api.allocate_for_instance( [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] created_port_ids = self._update_ports_for_instance( [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] with excutils.save_and_reraise_exception(): [ 701.431686] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] self.force_reraise() [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] raise self.value [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] updated_port = self._update_port( [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] _ensure_no_port_binding_failure(port) [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] raise exception.PortBindingFailed(port_id=port['id']) [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] nova.exception.PortBindingFailed: Binding failed for port 51b2e7bb-94cc-4ae3-a62a-e860d82899f9, please check neutron logs for more information. [ 701.432251] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] [ 701.432251] env[61857]: INFO nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Terminating instance [ 701.433103] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Acquiring lock "refresh_cache-952d0a1c-3e94-42dc-9bc4-8063275afc82" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.453144] env[61857]: DEBUG nova.scheduler.client.report [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.511285] env[61857]: INFO nova.compute.manager [-] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Took 1.02 seconds to deallocate network for instance. [ 701.633958] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.723680] env[61857]: DEBUG nova.network.neutron [req-e66a92f0-40fc-497f-ac33-8469d7b75dcc req-428fd910-b273-4533-a838-8afcfe0a8734 service nova] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.817298] env[61857]: DEBUG nova.network.neutron [req-e66a92f0-40fc-497f-ac33-8469d7b75dcc req-428fd910-b273-4533-a838-8afcfe0a8734 service nova] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.959211] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.959745] env[61857]: DEBUG nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.962618] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.313s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.963997] env[61857]: INFO nova.compute.claims [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 702.093373] env[61857]: INFO nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Took 0.58 seconds to detach 1 volumes for instance. [ 702.094455] env[61857]: DEBUG nova.compute.claims [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 702.094807] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.321793] env[61857]: DEBUG oslo_concurrency.lockutils [req-e66a92f0-40fc-497f-ac33-8469d7b75dcc req-428fd910-b273-4533-a838-8afcfe0a8734 service nova] Releasing lock "refresh_cache-952d0a1c-3e94-42dc-9bc4-8063275afc82" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.322228] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Acquired lock "refresh_cache-952d0a1c-3e94-42dc-9bc4-8063275afc82" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.322413] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 702.471767] env[61857]: DEBUG nova.compute.utils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 702.473206] env[61857]: DEBUG nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 702.473377] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 702.523561] env[61857]: DEBUG nova.policy [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a93a056bb9584561b78bf1ca50b2b763', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1fb6596a0004874b9ada4685011dbf1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 702.846136] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.855147] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Successfully created port: 99bdd524-9ab6-4c8a-8b03-dd16ea228712 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.957155] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.977409] env[61857]: DEBUG nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 703.259990] env[61857]: DEBUG nova.compute.manager [req-eaad925d-96f5-4f77-a76b-3887cf5cba29 req-aea23781-1e82-4a01-b5af-2583b00782da service nova] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Received event network-vif-deleted-51b2e7bb-94cc-4ae3-a62a-e860d82899f9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 703.393113] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfaaca6-c3a7-4e01-873b-0c68cdc2ca6d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.401029] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ded7e5e-31ab-496b-a732-7026eda49019 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.430847] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb69d07-0856-4883-ae8a-bb19bb507531 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.438200] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb01e68-f8b0-40e9-9228-a30aa0cc4e13 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.452585] env[61857]: DEBUG nova.compute.provider_tree [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.458462] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Releasing lock "refresh_cache-952d0a1c-3e94-42dc-9bc4-8063275afc82" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.459305] env[61857]: DEBUG nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.459305] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 703.459305] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c53325a-9a37-4ba1-ba69-95bbd81acbe0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.466918] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac28fe81-5f60-46a1-9e8a-313c73a5d62f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.488493] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 952d0a1c-3e94-42dc-9bc4-8063275afc82 could not be found. [ 703.488917] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 703.488917] env[61857]: INFO nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Took 0.03 seconds to destroy the instance on the hypervisor. [ 703.489133] env[61857]: DEBUG oslo.service.loopingcall [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.489558] env[61857]: DEBUG nova.compute.manager [-] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.489649] env[61857]: DEBUG nova.network.neutron [-] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 703.662066] env[61857]: DEBUG nova.network.neutron [-] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.956137] env[61857]: DEBUG nova.scheduler.client.report [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.992523] env[61857]: DEBUG nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 704.024786] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 704.024956] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 704.025130] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 704.025318] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 704.025464] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 704.025610] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 704.025811] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 704.025967] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 704.026462] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 704.026462] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 704.026573] env[61857]: DEBUG nova.virt.hardware [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.027527] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376ae8c0-e888-4f14-9fdb-9c9e2d9b7c76 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.035935] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a04728-41ef-444d-9999-234a77300971 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.120527] env[61857]: ERROR nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 99bdd524-9ab6-4c8a-8b03-dd16ea228712, please check neutron logs for more information. [ 704.120527] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 704.120527] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.120527] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 704.120527] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 704.120527] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 704.120527] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 704.120527] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 704.120527] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.120527] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 704.120527] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.120527] env[61857]: ERROR nova.compute.manager raise self.value [ 704.120527] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 704.120527] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 704.120527] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.120527] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 704.121011] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.121011] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 704.121011] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 99bdd524-9ab6-4c8a-8b03-dd16ea228712, please check neutron logs for more information. [ 704.121011] env[61857]: ERROR nova.compute.manager [ 704.121011] env[61857]: Traceback (most recent call last): [ 704.121011] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 704.121011] env[61857]: listener.cb(fileno) [ 704.121011] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.121011] env[61857]: result = function(*args, **kwargs) [ 704.121011] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 704.121011] env[61857]: return func(*args, **kwargs) [ 704.121011] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.121011] env[61857]: raise e [ 704.121011] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.121011] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 704.121011] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 704.121011] env[61857]: created_port_ids = self._update_ports_for_instance( [ 704.121011] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 704.121011] env[61857]: with excutils.save_and_reraise_exception(): [ 704.121011] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.121011] env[61857]: self.force_reraise() [ 704.121011] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.121011] env[61857]: raise self.value [ 704.121011] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 704.121011] env[61857]: updated_port = self._update_port( [ 704.121011] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.121011] env[61857]: _ensure_no_port_binding_failure(port) [ 704.121011] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.121011] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 704.121815] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 99bdd524-9ab6-4c8a-8b03-dd16ea228712, please check neutron logs for more information. [ 704.121815] env[61857]: Removing descriptor: 16 [ 704.121815] env[61857]: ERROR nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 99bdd524-9ab6-4c8a-8b03-dd16ea228712, please check neutron logs for more information. [ 704.121815] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Traceback (most recent call last): [ 704.121815] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 704.121815] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] yield resources [ 704.121815] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.121815] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] self.driver.spawn(context, instance, image_meta, [ 704.121815] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 704.121815] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.121815] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.121815] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] vm_ref = self.build_virtual_machine(instance, [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] for vif in network_info: [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] return self._sync_wrapper(fn, *args, **kwargs) [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] self.wait() [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] self[:] = self._gt.wait() [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] return self._exit_event.wait() [ 704.122164] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] result = hub.switch() [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] return self.greenlet.switch() [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] result = function(*args, **kwargs) [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] return func(*args, **kwargs) [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] raise e [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] nwinfo = self.network_api.allocate_for_instance( [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 704.122594] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] created_port_ids = self._update_ports_for_instance( [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] with excutils.save_and_reraise_exception(): [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] self.force_reraise() [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] raise self.value [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] updated_port = self._update_port( [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] _ensure_no_port_binding_failure(port) [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.122953] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] raise exception.PortBindingFailed(port_id=port['id']) [ 704.123325] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] nova.exception.PortBindingFailed: Binding failed for port 99bdd524-9ab6-4c8a-8b03-dd16ea228712, please check neutron logs for more information. [ 704.123325] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] [ 704.123325] env[61857]: INFO nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Terminating instance [ 704.123984] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Acquiring lock "refresh_cache-768e0e7d-7ebe-4d7a-a971-4fb463731050" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.124158] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Acquired lock "refresh_cache-768e0e7d-7ebe-4d7a-a971-4fb463731050" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.124324] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 704.164972] env[61857]: DEBUG nova.network.neutron [-] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.462580] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.463134] env[61857]: DEBUG nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.469372] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.655s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.646389] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.669278] env[61857]: INFO nova.compute.manager [-] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Took 1.18 seconds to deallocate network for instance. [ 704.670347] env[61857]: DEBUG nova.compute.claims [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.670347] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.744922] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.973562] env[61857]: DEBUG nova.compute.utils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.980105] env[61857]: DEBUG nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.980105] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 705.052264] env[61857]: DEBUG nova.policy [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9eee4d78af2540a8a2dfdcafa652a7ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f520132b3ea465aa25f665b8a7750bd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 705.252293] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Releasing lock "refresh_cache-768e0e7d-7ebe-4d7a-a971-4fb463731050" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.252764] env[61857]: DEBUG nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 705.252927] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 705.253216] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-813321cb-2d20-4317-89cc-7dddad16eda7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.263058] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7227b53d-06b9-401d-a08c-39b255be5dc2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.288519] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 768e0e7d-7ebe-4d7a-a971-4fb463731050 could not be found. [ 705.288674] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 705.288860] env[61857]: INFO nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Took 0.04 seconds to destroy the instance on the hypervisor. [ 705.289132] env[61857]: DEBUG oslo.service.loopingcall [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.291867] env[61857]: DEBUG nova.compute.manager [-] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.291970] env[61857]: DEBUG nova.network.neutron [-] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 705.309412] env[61857]: DEBUG nova.network.neutron [-] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.446701] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89666ea1-b42f-409a-8b70-427eb7968533 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.456757] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2248409b-0f87-4e60-9237-61df70396bc3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.462484] env[61857]: DEBUG nova.compute.manager [req-80b0eb33-326d-40ab-9290-df3fcf5fe507 req-3df6f25b-02aa-4bf0-84dd-ffd84dc593f5 service nova] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Received event network-changed-99bdd524-9ab6-4c8a-8b03-dd16ea228712 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 705.462863] env[61857]: DEBUG nova.compute.manager [req-80b0eb33-326d-40ab-9290-df3fcf5fe507 req-3df6f25b-02aa-4bf0-84dd-ffd84dc593f5 service nova] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Refreshing instance network info cache due to event network-changed-99bdd524-9ab6-4c8a-8b03-dd16ea228712. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 705.463116] env[61857]: DEBUG oslo_concurrency.lockutils [req-80b0eb33-326d-40ab-9290-df3fcf5fe507 req-3df6f25b-02aa-4bf0-84dd-ffd84dc593f5 service nova] Acquiring lock "refresh_cache-768e0e7d-7ebe-4d7a-a971-4fb463731050" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.463269] env[61857]: DEBUG oslo_concurrency.lockutils [req-80b0eb33-326d-40ab-9290-df3fcf5fe507 req-3df6f25b-02aa-4bf0-84dd-ffd84dc593f5 service nova] Acquired lock "refresh_cache-768e0e7d-7ebe-4d7a-a971-4fb463731050" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.463564] env[61857]: DEBUG nova.network.neutron [req-80b0eb33-326d-40ab-9290-df3fcf5fe507 req-3df6f25b-02aa-4bf0-84dd-ffd84dc593f5 service nova] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Refreshing network info cache for port 99bdd524-9ab6-4c8a-8b03-dd16ea228712 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 705.494387] env[61857]: DEBUG nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.499018] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222ce431-dd63-4f0b-a07d-adba90c8ce70 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.508707] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a9ea33-9e11-4107-bb35-5d9726d2e9da {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.527181] env[61857]: DEBUG nova.compute.provider_tree [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.557180] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Successfully created port: c9b09025-f020-4647-9acb-5be89eb1ac86 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.813414] env[61857]: DEBUG nova.network.neutron [-] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.869810] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Successfully created port: d1009629-3b35-4af8-94ce-bbfe827d0549 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.985053] env[61857]: DEBUG nova.network.neutron [req-80b0eb33-326d-40ab-9290-df3fcf5fe507 req-3df6f25b-02aa-4bf0-84dd-ffd84dc593f5 service nova] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.029968] env[61857]: DEBUG nova.scheduler.client.report [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.116871] env[61857]: DEBUG nova.network.neutron [req-80b0eb33-326d-40ab-9290-df3fcf5fe507 req-3df6f25b-02aa-4bf0-84dd-ffd84dc593f5 service nova] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.282874] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Successfully created port: 165dda43-44e8-43cd-803a-155bb3330abd {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.314564] env[61857]: INFO nova.compute.manager [-] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Took 1.02 seconds to deallocate network for instance. [ 706.320540] env[61857]: DEBUG nova.compute.claims [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 706.320540] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.514210] env[61857]: DEBUG nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.536108] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.070s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.536842] env[61857]: ERROR nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fe95c043-c85d-470c-ac22-755174721da1, please check neutron logs for more information. [ 706.536842] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Traceback (most recent call last): [ 706.536842] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.536842] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] self.driver.spawn(context, instance, image_meta, [ 706.536842] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 706.536842] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.536842] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.536842] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] vm_ref = self.build_virtual_machine(instance, [ 706.536842] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.536842] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.536842] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] for vif in network_info: [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] return self._sync_wrapper(fn, *args, **kwargs) [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] self.wait() [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] self[:] = self._gt.wait() [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] return self._exit_event.wait() [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] result = hub.switch() [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 706.537227] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] return self.greenlet.switch() [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] result = function(*args, **kwargs) [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] return func(*args, **kwargs) [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] raise e [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] nwinfo = self.network_api.allocate_for_instance( [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] created_port_ids = self._update_ports_for_instance( [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] with excutils.save_and_reraise_exception(): [ 706.537607] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] self.force_reraise() [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] raise self.value [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] updated_port = self._update_port( [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] _ensure_no_port_binding_failure(port) [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] raise exception.PortBindingFailed(port_id=port['id']) [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] nova.exception.PortBindingFailed: Binding failed for port fe95c043-c85d-470c-ac22-755174721da1, please check neutron logs for more information. [ 706.537930] env[61857]: ERROR nova.compute.manager [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] [ 706.538221] env[61857]: DEBUG nova.compute.utils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Binding failed for port fe95c043-c85d-470c-ac22-755174721da1, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 706.539609] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.177s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.541241] env[61857]: INFO nova.compute.claims [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.548742] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Build of instance 51c54fe3-c0e2-4151-9626-cfb10cd7996e was re-scheduled: Binding failed for port fe95c043-c85d-470c-ac22-755174721da1, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 706.549992] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 706.550326] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "refresh_cache-51c54fe3-c0e2-4151-9626-cfb10cd7996e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.550990] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired lock "refresh_cache-51c54fe3-c0e2-4151-9626-cfb10cd7996e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.551244] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.556178] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.556401] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.556603] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.556815] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.556962] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.557123] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.557329] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.557485] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.557647] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.557806] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.557977] env[61857]: DEBUG nova.virt.hardware [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.559077] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867685cc-4ac0-4443-aeb4-f29c1fff4414 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.568158] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e115780-bc5a-47ac-9f83-1baf27654450 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.620667] env[61857]: DEBUG oslo_concurrency.lockutils [req-80b0eb33-326d-40ab-9290-df3fcf5fe507 req-3df6f25b-02aa-4bf0-84dd-ffd84dc593f5 service nova] Releasing lock "refresh_cache-768e0e7d-7ebe-4d7a-a971-4fb463731050" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.620942] env[61857]: DEBUG nova.compute.manager [req-80b0eb33-326d-40ab-9290-df3fcf5fe507 req-3df6f25b-02aa-4bf0-84dd-ffd84dc593f5 service nova] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Received event network-vif-deleted-99bdd524-9ab6-4c8a-8b03-dd16ea228712 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 706.990052] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "93e03e35-15a0-49e7-b1be-09178eabbfda" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.990257] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "93e03e35-15a0-49e7-b1be-09178eabbfda" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.079261] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.155716] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.402777] env[61857]: ERROR nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c9b09025-f020-4647-9acb-5be89eb1ac86, please check neutron logs for more information. [ 707.402777] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 707.402777] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.402777] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 707.402777] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.402777] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 707.402777] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.402777] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 707.402777] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.402777] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 707.402777] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.402777] env[61857]: ERROR nova.compute.manager raise self.value [ 707.402777] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.402777] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 707.402777] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.402777] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 707.403294] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.403294] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 707.403294] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c9b09025-f020-4647-9acb-5be89eb1ac86, please check neutron logs for more information. [ 707.403294] env[61857]: ERROR nova.compute.manager [ 707.403294] env[61857]: Traceback (most recent call last): [ 707.403294] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 707.403294] env[61857]: listener.cb(fileno) [ 707.403294] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.403294] env[61857]: result = function(*args, **kwargs) [ 707.403294] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.403294] env[61857]: return func(*args, **kwargs) [ 707.403294] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.403294] env[61857]: raise e [ 707.403294] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.403294] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 707.403294] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.403294] env[61857]: created_port_ids = self._update_ports_for_instance( [ 707.403294] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.403294] env[61857]: with excutils.save_and_reraise_exception(): [ 707.403294] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.403294] env[61857]: self.force_reraise() [ 707.403294] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.403294] env[61857]: raise self.value [ 707.403294] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.403294] env[61857]: updated_port = self._update_port( [ 707.403294] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.403294] env[61857]: _ensure_no_port_binding_failure(port) [ 707.403294] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.403294] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 707.404229] env[61857]: nova.exception.PortBindingFailed: Binding failed for port c9b09025-f020-4647-9acb-5be89eb1ac86, please check neutron logs for more information. [ 707.404229] env[61857]: Removing descriptor: 16 [ 707.404229] env[61857]: ERROR nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c9b09025-f020-4647-9acb-5be89eb1ac86, please check neutron logs for more information. [ 707.404229] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Traceback (most recent call last): [ 707.404229] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 707.404229] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] yield resources [ 707.404229] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.404229] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] self.driver.spawn(context, instance, image_meta, [ 707.404229] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 707.404229] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.404229] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.404229] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] vm_ref = self.build_virtual_machine(instance, [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] for vif in network_info: [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] return self._sync_wrapper(fn, *args, **kwargs) [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] self.wait() [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] self[:] = self._gt.wait() [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] return self._exit_event.wait() [ 707.404588] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] result = hub.switch() [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] return self.greenlet.switch() [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] result = function(*args, **kwargs) [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] return func(*args, **kwargs) [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] raise e [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] nwinfo = self.network_api.allocate_for_instance( [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 707.404987] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] created_port_ids = self._update_ports_for_instance( [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] with excutils.save_and_reraise_exception(): [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] self.force_reraise() [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] raise self.value [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] updated_port = self._update_port( [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] _ensure_no_port_binding_failure(port) [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.405394] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] raise exception.PortBindingFailed(port_id=port['id']) [ 707.405766] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] nova.exception.PortBindingFailed: Binding failed for port c9b09025-f020-4647-9acb-5be89eb1ac86, please check neutron logs for more information. [ 707.405766] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] [ 707.405766] env[61857]: INFO nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Terminating instance [ 707.409488] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "refresh_cache-b3fd3980-efe4-49a0-b656-32f1543e657c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.409655] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquired lock "refresh_cache-b3fd3980-efe4-49a0-b656-32f1543e657c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.409829] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 707.486591] env[61857]: DEBUG nova.compute.manager [req-dd0ae84f-8f4d-4325-ae2a-319badff632f req-455cb545-95af-4e06-b66a-4966bd9ec3c4 service nova] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Received event network-changed-c9b09025-f020-4647-9acb-5be89eb1ac86 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.486865] env[61857]: DEBUG nova.compute.manager [req-dd0ae84f-8f4d-4325-ae2a-319badff632f req-455cb545-95af-4e06-b66a-4966bd9ec3c4 service nova] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Refreshing instance network info cache due to event network-changed-c9b09025-f020-4647-9acb-5be89eb1ac86. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 707.487083] env[61857]: DEBUG oslo_concurrency.lockutils [req-dd0ae84f-8f4d-4325-ae2a-319badff632f req-455cb545-95af-4e06-b66a-4966bd9ec3c4 service nova] Acquiring lock "refresh_cache-b3fd3980-efe4-49a0-b656-32f1543e657c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.659204] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Releasing lock "refresh_cache-51c54fe3-c0e2-4151-9626-cfb10cd7996e" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.659204] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 707.659204] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.659204] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 707.674366] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.953018] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fbcb6ec-d143-4eee-adca-78521bfd66cb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.964763] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fefa28d6-2218-4194-9fc8-922dbe7730cf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.968662] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.999333] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2cebda9-f08f-474b-a3ca-9070ccf7b525 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.007531] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc42274-15da-4b15-8132-29d16506255f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.021460] env[61857]: DEBUG nova.compute.provider_tree [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.097688] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.178508] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.526085] env[61857]: DEBUG nova.scheduler.client.report [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.602483] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Releasing lock "refresh_cache-b3fd3980-efe4-49a0-b656-32f1543e657c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.602483] env[61857]: DEBUG nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 708.602483] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 708.602483] env[61857]: DEBUG oslo_concurrency.lockutils [req-dd0ae84f-8f4d-4325-ae2a-319badff632f req-455cb545-95af-4e06-b66a-4966bd9ec3c4 service nova] Acquired lock "refresh_cache-b3fd3980-efe4-49a0-b656-32f1543e657c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.602483] env[61857]: DEBUG nova.network.neutron [req-dd0ae84f-8f4d-4325-ae2a-319badff632f req-455cb545-95af-4e06-b66a-4966bd9ec3c4 service nova] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Refreshing network info cache for port c9b09025-f020-4647-9acb-5be89eb1ac86 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 708.602707] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7f73fa7-1222-4930-b6c9-013f289587e7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.612797] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506d74ff-2471-42c1-8372-6bb8a8141da1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.637730] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b3fd3980-efe4-49a0-b656-32f1543e657c could not be found. [ 708.637954] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 708.638174] env[61857]: INFO nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 708.638428] env[61857]: DEBUG oslo.service.loopingcall [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.638641] env[61857]: DEBUG nova.compute.manager [-] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.638736] env[61857]: DEBUG nova.network.neutron [-] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 708.671237] env[61857]: DEBUG nova.network.neutron [-] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.680601] env[61857]: INFO nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 51c54fe3-c0e2-4151-9626-cfb10cd7996e] Took 1.02 seconds to deallocate network for instance. [ 709.034800] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.034800] env[61857]: DEBUG nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 709.035273] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.530s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.130862] env[61857]: DEBUG nova.network.neutron [req-dd0ae84f-8f4d-4325-ae2a-319badff632f req-455cb545-95af-4e06-b66a-4966bd9ec3c4 service nova] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.213888] env[61857]: DEBUG nova.network.neutron [req-dd0ae84f-8f4d-4325-ae2a-319badff632f req-455cb545-95af-4e06-b66a-4966bd9ec3c4 service nova] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.540899] env[61857]: DEBUG nova.compute.utils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.545563] env[61857]: DEBUG nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 709.545738] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 709.589875] env[61857]: DEBUG nova.policy [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1a9fb89bc8d4c7cbc14b23dee76806b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69e69c651a7f4309a808ed798dd87343', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 709.623833] env[61857]: DEBUG nova.network.neutron [-] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.715989] env[61857]: DEBUG oslo_concurrency.lockutils [req-dd0ae84f-8f4d-4325-ae2a-319badff632f req-455cb545-95af-4e06-b66a-4966bd9ec3c4 service nova] Releasing lock "refresh_cache-b3fd3980-efe4-49a0-b656-32f1543e657c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.716375] env[61857]: DEBUG nova.compute.manager [req-dd0ae84f-8f4d-4325-ae2a-319badff632f req-455cb545-95af-4e06-b66a-4966bd9ec3c4 service nova] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Received event network-vif-deleted-c9b09025-f020-4647-9acb-5be89eb1ac86 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 709.718243] env[61857]: INFO nova.scheduler.client.report [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Deleted allocations for instance 51c54fe3-c0e2-4151-9626-cfb10cd7996e [ 709.937852] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Successfully created port: 619c90fa-1887-4f9c-b8a7-ad58472dbbf5 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.977028] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5b7df5-9390-4ceb-84c8-c8aeac48c5d0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.984141] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12caf8f6-8a5d-4e5e-9bd4-145781a55606 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.017587] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3abad92-cbf9-4237-8bb1-058a6f1ea531 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.024699] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b4782a-3553-4caf-9e14-2c0363dc88ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.037841] env[61857]: DEBUG nova.compute.provider_tree [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.045952] env[61857]: DEBUG nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.125682] env[61857]: INFO nova.compute.manager [-] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Took 1.49 seconds to deallocate network for instance. [ 710.128894] env[61857]: DEBUG nova.compute.claims [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 710.129069] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.226757] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "51c54fe3-c0e2-4151-9626-cfb10cd7996e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.910s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.542139] env[61857]: DEBUG nova.scheduler.client.report [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.729495] env[61857]: DEBUG nova.compute.manager [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 710.814102] env[61857]: DEBUG nova.compute.manager [req-e14764b9-b5bd-49f6-9649-fef7c89184de req-a412d099-4704-4f86-84df-3e5beb402e1a service nova] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Received event network-changed-619c90fa-1887-4f9c-b8a7-ad58472dbbf5 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 710.814315] env[61857]: DEBUG nova.compute.manager [req-e14764b9-b5bd-49f6-9649-fef7c89184de req-a412d099-4704-4f86-84df-3e5beb402e1a service nova] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Refreshing instance network info cache due to event network-changed-619c90fa-1887-4f9c-b8a7-ad58472dbbf5. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 710.814891] env[61857]: DEBUG oslo_concurrency.lockutils [req-e14764b9-b5bd-49f6-9649-fef7c89184de req-a412d099-4704-4f86-84df-3e5beb402e1a service nova] Acquiring lock "refresh_cache-e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.815230] env[61857]: DEBUG oslo_concurrency.lockutils [req-e14764b9-b5bd-49f6-9649-fef7c89184de req-a412d099-4704-4f86-84df-3e5beb402e1a service nova] Acquired lock "refresh_cache-e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.820286] env[61857]: DEBUG nova.network.neutron [req-e14764b9-b5bd-49f6-9649-fef7c89184de req-a412d099-4704-4f86-84df-3e5beb402e1a service nova] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Refreshing network info cache for port 619c90fa-1887-4f9c-b8a7-ad58472dbbf5 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 710.922087] env[61857]: ERROR nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 619c90fa-1887-4f9c-b8a7-ad58472dbbf5, please check neutron logs for more information. [ 710.922087] env[61857]: ERROR nova.compute.manager Traceback (most recent call last): [ 710.922087] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.922087] env[61857]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 710.922087] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 710.922087] env[61857]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 710.922087] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 710.922087] env[61857]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 710.922087] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.922087] env[61857]: ERROR nova.compute.manager self.force_reraise() [ 710.922087] env[61857]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.922087] env[61857]: ERROR nova.compute.manager raise self.value [ 710.922087] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 710.922087] env[61857]: ERROR nova.compute.manager updated_port = self._update_port( [ 710.922087] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.922087] env[61857]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 710.922778] env[61857]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.922778] env[61857]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 710.922778] env[61857]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 619c90fa-1887-4f9c-b8a7-ad58472dbbf5, please check neutron logs for more information. [ 710.922778] env[61857]: ERROR nova.compute.manager [ 710.922778] env[61857]: Traceback (most recent call last): [ 710.922778] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 710.922778] env[61857]: listener.cb(fileno) [ 710.922778] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.922778] env[61857]: result = function(*args, **kwargs) [ 710.922778] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.922778] env[61857]: return func(*args, **kwargs) [ 710.922778] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.922778] env[61857]: raise e [ 710.922778] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.922778] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 710.922778] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 710.922778] env[61857]: created_port_ids = self._update_ports_for_instance( [ 710.922778] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 710.922778] env[61857]: with excutils.save_and_reraise_exception(): [ 710.922778] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.922778] env[61857]: self.force_reraise() [ 710.922778] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.922778] env[61857]: raise self.value [ 710.922778] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 710.922778] env[61857]: updated_port = self._update_port( [ 710.922778] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.922778] env[61857]: _ensure_no_port_binding_failure(port) [ 710.922778] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.922778] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 710.923544] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 619c90fa-1887-4f9c-b8a7-ad58472dbbf5, please check neutron logs for more information. [ 710.923544] env[61857]: Removing descriptor: 17 [ 711.049278] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.014s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.052677] env[61857]: ERROR nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. [ 711.052677] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Traceback (most recent call last): [ 711.052677] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.052677] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] self.driver.spawn(context, instance, image_meta, [ 711.052677] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 711.052677] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.052677] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.052677] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] vm_ref = self.build_virtual_machine(instance, [ 711.052677] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.052677] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.052677] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] for vif in network_info: [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] return self._sync_wrapper(fn, *args, **kwargs) [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] self.wait() [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] self[:] = self._gt.wait() [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] return self._exit_event.wait() [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] result = hub.switch() [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 711.053267] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] return self.greenlet.switch() [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] result = function(*args, **kwargs) [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] return func(*args, **kwargs) [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] raise e [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] nwinfo = self.network_api.allocate_for_instance( [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] created_port_ids = self._update_ports_for_instance( [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] with excutils.save_and_reraise_exception(): [ 711.053757] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] self.force_reraise() [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] raise self.value [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] updated_port = self._update_port( [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] _ensure_no_port_binding_failure(port) [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] raise exception.PortBindingFailed(port_id=port['id']) [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] nova.exception.PortBindingFailed: Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. [ 711.054248] env[61857]: ERROR nova.compute.manager [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] [ 711.054948] env[61857]: DEBUG nova.compute.utils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 711.056977] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.715s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.060888] env[61857]: DEBUG nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 711.063248] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Build of instance 8ee206cf-c906-4ac0-8ba7-d864f5f538db was re-scheduled: Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 711.064773] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 711.064773] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "refresh_cache-8ee206cf-c906-4ac0-8ba7-d864f5f538db" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.064893] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired lock "refresh_cache-8ee206cf-c906-4ac0-8ba7-d864f5f538db" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.065050] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 711.090664] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.090918] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.091224] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.092016] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.092016] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.092016] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.092016] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.092203] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.092249] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.092408] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.092792] env[61857]: DEBUG nova.virt.hardware [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.093962] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65df839a-8cb3-436d-8af5-448755a3f620 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.106796] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13641edc-7914-4437-9db3-185079cd115c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.125147] env[61857]: ERROR nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 619c90fa-1887-4f9c-b8a7-ad58472dbbf5, please check neutron logs for more information. [ 711.125147] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Traceback (most recent call last): [ 711.125147] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 711.125147] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] yield resources [ 711.125147] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.125147] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] self.driver.spawn(context, instance, image_meta, [ 711.125147] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 711.125147] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.125147] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.125147] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] vm_ref = self.build_virtual_machine(instance, [ 711.125147] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] for vif in network_info: [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] return self._sync_wrapper(fn, *args, **kwargs) [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] self.wait() [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] self[:] = self._gt.wait() [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] return self._exit_event.wait() [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 711.125615] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] current.throw(*self._exc) [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] result = function(*args, **kwargs) [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] return func(*args, **kwargs) [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] raise e [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] nwinfo = self.network_api.allocate_for_instance( [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] created_port_ids = self._update_ports_for_instance( [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] with excutils.save_and_reraise_exception(): [ 711.125979] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] self.force_reraise() [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] raise self.value [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] updated_port = self._update_port( [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] _ensure_no_port_binding_failure(port) [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] raise exception.PortBindingFailed(port_id=port['id']) [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] nova.exception.PortBindingFailed: Binding failed for port 619c90fa-1887-4f9c-b8a7-ad58472dbbf5, please check neutron logs for more information. [ 711.126355] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] [ 711.126355] env[61857]: INFO nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Terminating instance [ 711.128212] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Acquiring lock "refresh_cache-e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.256572] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.359873] env[61857]: DEBUG nova.network.neutron [req-e14764b9-b5bd-49f6-9649-fef7c89184de req-a412d099-4704-4f86-84df-3e5beb402e1a service nova] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.476163] env[61857]: DEBUG nova.network.neutron [req-e14764b9-b5bd-49f6-9649-fef7c89184de req-a412d099-4704-4f86-84df-3e5beb402e1a service nova] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.569074] env[61857]: DEBUG nova.compute.utils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Can not refresh info_cache because instance was not found {{(pid=61857) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 711.574047] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.574292] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.588581] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.607972] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "a73f3163-b6fc-4764-b76a-451f823b5808" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.608245] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "a73f3163-b6fc-4764-b76a-451f823b5808" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.661944] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.911566] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913031fa-f009-4939-bf22-486696127e6e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.919143] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-110f4ab7-fe52-4c0f-81f3-8474134b4d72 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.951187] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d023751-9912-4675-aab4-31074606a127 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.958514] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726d1782-3c96-45c8-874e-c03a7f639ba8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.972026] env[61857]: DEBUG nova.compute.provider_tree [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.979687] env[61857]: DEBUG oslo_concurrency.lockutils [req-e14764b9-b5bd-49f6-9649-fef7c89184de req-a412d099-4704-4f86-84df-3e5beb402e1a service nova] Releasing lock "refresh_cache-e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.979889] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Acquired lock "refresh_cache-e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.980072] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.165058] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Releasing lock "refresh_cache-8ee206cf-c906-4ac0-8ba7-d864f5f538db" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.165058] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 712.165058] env[61857]: DEBUG nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.165058] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 712.180041] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.474456] env[61857]: DEBUG nova.scheduler.client.report [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.501907] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.600516] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.683212] env[61857]: DEBUG nova.network.neutron [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.839930] env[61857]: DEBUG nova.compute.manager [req-532c7396-e669-470c-bdd3-cb199b2243fd req-7a284cd9-e044-4a54-a97b-ec37562a8a8d service nova] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Received event network-vif-deleted-619c90fa-1887-4f9c-b8a7-ad58472dbbf5 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.979787] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.923s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.980597] env[61857]: ERROR nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6, please check neutron logs for more information. [ 712.980597] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Traceback (most recent call last): [ 712.980597] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.980597] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] self.driver.spawn(context, instance, image_meta, [ 712.980597] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 712.980597] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.980597] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.980597] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] vm_ref = self.build_virtual_machine(instance, [ 712.980597] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.980597] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.980597] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] for vif in network_info: [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] return self._sync_wrapper(fn, *args, **kwargs) [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] self.wait() [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] self[:] = self._gt.wait() [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] return self._exit_event.wait() [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] result = hub.switch() [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.980931] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] return self.greenlet.switch() [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] result = function(*args, **kwargs) [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] return func(*args, **kwargs) [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] raise e [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] nwinfo = self.network_api.allocate_for_instance( [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] created_port_ids = self._update_ports_for_instance( [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] with excutils.save_and_reraise_exception(): [ 712.981295] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] self.force_reraise() [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] raise self.value [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] updated_port = self._update_port( [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] _ensure_no_port_binding_failure(port) [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] raise exception.PortBindingFailed(port_id=port['id']) [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] nova.exception.PortBindingFailed: Binding failed for port 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6, please check neutron logs for more information. [ 712.981725] env[61857]: ERROR nova.compute.manager [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] [ 712.982021] env[61857]: DEBUG nova.compute.utils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Binding failed for port 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.982731] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.345s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.986013] env[61857]: DEBUG nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Build of instance ca3d3ad9-39a0-467b-812f-21b7303b8bc8 was re-scheduled: Binding failed for port 967189fc-2ecf-4f90-ab01-8af7ce3fe2c6, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.986431] env[61857]: DEBUG nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.986655] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "refresh_cache-ca3d3ad9-39a0-467b-812f-21b7303b8bc8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.986801] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "refresh_cache-ca3d3ad9-39a0-467b-812f-21b7303b8bc8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.986975] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 713.103627] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Releasing lock "refresh_cache-e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.104044] env[61857]: DEBUG nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 713.104277] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 713.104572] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ba73930-f282-4891-b2e5-d162783bda2d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.113405] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba6237b-41b8-4d9a-a28d-f2172ea47e3a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.135304] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd could not be found. [ 713.135535] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 713.135720] env[61857]: INFO nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 713.135960] env[61857]: DEBUG oslo.service.loopingcall [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.136200] env[61857]: DEBUG nova.compute.manager [-] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.136530] env[61857]: DEBUG nova.network.neutron [-] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 713.152974] env[61857]: DEBUG nova.network.neutron [-] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.185390] env[61857]: INFO nova.compute.manager [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: 8ee206cf-c906-4ac0-8ba7-d864f5f538db] Took 1.02 seconds to deallocate network for instance. [ 713.510848] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.600013] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.655815] env[61857]: DEBUG nova.network.neutron [-] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.845331] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e6521a-5939-4e0a-88b6-2d9bec3c6272 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.852599] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3fde10-172c-4e01-b54b-652391dbeb31 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.881765] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac59cff7-39ca-4187-ad50-2bdd0f7bf7ad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.888736] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8e6916-d8f7-4123-867b-fe3d05c96f2e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.901609] env[61857]: DEBUG nova.compute.provider_tree [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.102991] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "refresh_cache-ca3d3ad9-39a0-467b-812f-21b7303b8bc8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.103410] env[61857]: DEBUG nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 714.103460] env[61857]: DEBUG nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.103626] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 714.119433] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.158033] env[61857]: INFO nova.compute.manager [-] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Took 1.02 seconds to deallocate network for instance. [ 714.160423] env[61857]: DEBUG nova.compute.claims [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Aborting claim: {{(pid=61857) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 714.160601] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.404995] env[61857]: DEBUG nova.scheduler.client.report [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.622758] env[61857]: DEBUG nova.network.neutron [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.910450] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.928s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.911147] env[61857]: ERROR nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 815a603a-b57e-4bca-823f-635e83967652, please check neutron logs for more information. [ 714.911147] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Traceback (most recent call last): [ 714.911147] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.911147] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] self.driver.spawn(context, instance, image_meta, [ 714.911147] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 714.911147] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.911147] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.911147] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] vm_ref = self.build_virtual_machine(instance, [ 714.911147] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.911147] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.911147] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] for vif in network_info: [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] return self._sync_wrapper(fn, *args, **kwargs) [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] self.wait() [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] self[:] = self._gt.wait() [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] return self._exit_event.wait() [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] result = hub.switch() [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 714.911476] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] return self.greenlet.switch() [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] result = function(*args, **kwargs) [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] return func(*args, **kwargs) [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] raise e [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] nwinfo = self.network_api.allocate_for_instance( [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] created_port_ids = self._update_ports_for_instance( [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] with excutils.save_and_reraise_exception(): [ 714.911802] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] self.force_reraise() [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] raise self.value [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] updated_port = self._update_port( [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] _ensure_no_port_binding_failure(port) [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] raise exception.PortBindingFailed(port_id=port['id']) [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] nova.exception.PortBindingFailed: Binding failed for port 815a603a-b57e-4bca-823f-635e83967652, please check neutron logs for more information. [ 714.912143] env[61857]: ERROR nova.compute.manager [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] [ 714.912462] env[61857]: DEBUG nova.compute.utils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Binding failed for port 815a603a-b57e-4bca-823f-635e83967652, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 714.913167] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.279s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.914729] env[61857]: INFO nova.compute.claims [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.917413] env[61857]: DEBUG nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Build of instance 9edf9e8c-5421-4018-afee-59e10cd335ef was re-scheduled: Binding failed for port 815a603a-b57e-4bca-823f-635e83967652, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 714.917834] env[61857]: DEBUG nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 714.918074] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "refresh_cache-9edf9e8c-5421-4018-afee-59e10cd335ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.918243] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquired lock "refresh_cache-9edf9e8c-5421-4018-afee-59e10cd335ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.918424] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 715.125618] env[61857]: INFO nova.compute.manager [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ca3d3ad9-39a0-467b-812f-21b7303b8bc8] Took 1.02 seconds to deallocate network for instance. [ 715.219790] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c2beef0f-0374-488b-bbed-4d19a3d1dfc5 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "8ee206cf-c906-4ac0-8ba7-d864f5f538db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.865s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.220504] env[61857]: Traceback (most recent call last): [ 715.220596] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.220596] env[61857]: self.driver.spawn(context, instance, image_meta, [ 715.220596] env[61857]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 715.220596] env[61857]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.220596] env[61857]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.220596] env[61857]: vm_ref = self.build_virtual_machine(instance, [ 715.220596] env[61857]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.220596] env[61857]: vif_infos = vmwarevif.get_vif_info(self._session, [ 715.220596] env[61857]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.220596] env[61857]: for vif in network_info: [ 715.220886] env[61857]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.220886] env[61857]: return self._sync_wrapper(fn, *args, **kwargs) [ 715.220886] env[61857]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.220886] env[61857]: self.wait() [ 715.220886] env[61857]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.220886] env[61857]: self[:] = self._gt.wait() [ 715.220886] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.220886] env[61857]: return self._exit_event.wait() [ 715.220886] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 715.220886] env[61857]: result = hub.switch() [ 715.220886] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 715.220886] env[61857]: return self.greenlet.switch() [ 715.220886] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.220886] env[61857]: result = function(*args, **kwargs) [ 715.220886] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.220886] env[61857]: return func(*args, **kwargs) [ 715.220886] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.220886] env[61857]: raise e [ 715.220886] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.220886] env[61857]: nwinfo = self.network_api.allocate_for_instance( [ 715.220886] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 715.220886] env[61857]: created_port_ids = self._update_ports_for_instance( [ 715.220886] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 715.220886] env[61857]: with excutils.save_and_reraise_exception(): [ 715.220886] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.220886] env[61857]: self.force_reraise() [ 715.220886] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.220886] env[61857]: raise self.value [ 715.220886] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 715.220886] env[61857]: updated_port = self._update_port( [ 715.220886] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.220886] env[61857]: _ensure_no_port_binding_failure(port) [ 715.221780] env[61857]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.221780] env[61857]: raise exception.PortBindingFailed(port_id=port['id']) [ 715.221780] env[61857]: nova.exception.PortBindingFailed: Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. [ 715.221780] env[61857]: During handling of the above exception, another exception occurred: [ 715.221780] env[61857]: Traceback (most recent call last): [ 715.221780] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 715.221780] env[61857]: self._build_and_run_instance(context, instance, image, [ 715.221780] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2748, in _build_and_run_instance [ 715.221780] env[61857]: raise exception.RescheduledException( [ 715.221780] env[61857]: nova.exception.RescheduledException: Build of instance 8ee206cf-c906-4ac0-8ba7-d864f5f538db was re-scheduled: Binding failed for port 9e56e450-a058-4ede-b734-ca42739a4736, please check neutron logs for more information. [ 715.221780] env[61857]: During handling of the above exception, another exception occurred: [ 715.221780] env[61857]: Traceback (most recent call last): [ 715.221780] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 715.221780] env[61857]: func(*args, **kwargs) [ 715.221780] env[61857]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.221780] env[61857]: return func(*args, **kwargs) [ 715.221780] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 715.221780] env[61857]: return f(*args, **kwargs) [ 715.221780] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 715.221780] env[61857]: result = self._do_build_and_run_instance(*args, **kwargs) [ 715.221780] env[61857]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 715.221780] env[61857]: with excutils.save_and_reraise_exception(): [ 715.221780] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.221780] env[61857]: self.force_reraise() [ 715.221780] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.221780] env[61857]: raise self.value [ 715.221780] env[61857]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 715.221780] env[61857]: return f(self, context, *args, **kw) [ 715.222596] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 715.222596] env[61857]: with excutils.save_and_reraise_exception(): [ 715.222596] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.222596] env[61857]: self.force_reraise() [ 715.222596] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.222596] env[61857]: raise self.value [ 715.222596] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 715.222596] env[61857]: return function(self, context, *args, **kwargs) [ 715.222596] env[61857]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 715.222596] env[61857]: return function(self, context, *args, **kwargs) [ 715.222596] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 715.222596] env[61857]: return function(self, context, *args, **kwargs) [ 715.222596] env[61857]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 715.222596] env[61857]: instance.save() [ 715.222596] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 715.222596] env[61857]: updates, result = self.indirection_api.object_action( [ 715.222596] env[61857]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 715.222596] env[61857]: return cctxt.call(context, 'object_action', objinst=objinst, [ 715.222596] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 715.222596] env[61857]: result = self.transport._send( [ 715.222596] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 715.222596] env[61857]: return self._driver.send(target, ctxt, message, [ 715.222596] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 715.222596] env[61857]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 715.222596] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 715.222596] env[61857]: raise result [ 715.222596] env[61857]: nova.exception_Remote.InstanceNotFound_Remote: Instance 8ee206cf-c906-4ac0-8ba7-d864f5f538db could not be found. [ 715.222596] env[61857]: Traceback (most recent call last): [ 715.223516] env[61857]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 715.223516] env[61857]: return getattr(target, method)(*args, **kwargs) [ 715.223516] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 715.223516] env[61857]: return fn(self, *args, **kwargs) [ 715.223516] env[61857]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 715.223516] env[61857]: old_ref, inst_ref = db.instance_update_and_get_original( [ 715.223516] env[61857]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 715.223516] env[61857]: return f(*args, **kwargs) [ 715.223516] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 715.223516] env[61857]: with excutils.save_and_reraise_exception() as ectxt: [ 715.223516] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.223516] env[61857]: self.force_reraise() [ 715.223516] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.223516] env[61857]: raise self.value [ 715.223516] env[61857]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 715.223516] env[61857]: return f(*args, **kwargs) [ 715.223516] env[61857]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 715.223516] env[61857]: return f(context, *args, **kwargs) [ 715.223516] env[61857]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 715.223516] env[61857]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 715.223516] env[61857]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 715.223516] env[61857]: raise exception.InstanceNotFound(instance_id=uuid) [ 715.223516] env[61857]: nova.exception.InstanceNotFound: Instance 8ee206cf-c906-4ac0-8ba7-d864f5f538db could not be found. [ 715.439938] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.507961] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.724506] env[61857]: DEBUG nova.compute.manager [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.010132] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Releasing lock "refresh_cache-9edf9e8c-5421-4018-afee-59e10cd335ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.010867] env[61857]: DEBUG nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 716.010867] env[61857]: DEBUG nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.010867] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 716.025205] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.151261] env[61857]: INFO nova.scheduler.client.report [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleted allocations for instance ca3d3ad9-39a0-467b-812f-21b7303b8bc8 [ 716.243785] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.266914] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bae53d-dbc9-4eee-8c9e-b53beadca219 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.277019] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3300a291-d0c3-49a7-b54c-75906d8f517f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.305828] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623d3820-a298-420a-aa6d-8b7c5b6fd9b0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.313390] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6710c66-ec3e-4917-9757-5675aadceb08 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.326525] env[61857]: DEBUG nova.compute.provider_tree [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.344123] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.344123] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.344123] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 716.344287] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Rebuilding the list of instances to heal {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 716.528510] env[61857]: DEBUG nova.network.neutron [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.662417] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fb7d8d9c-f558-4d34-beb8-2d1bb2b0a757 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "ca3d3ad9-39a0-467b-812f-21b7303b8bc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.197s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.829985] env[61857]: DEBUG nova.scheduler.client.report [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.847765] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 716.848063] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 716.848237] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 716.848373] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 716.848499] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 716.848624] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 716.848746] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Didn't find any instances for network info cache update. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 716.849685] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.849685] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.849685] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.849685] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.849933] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.849933] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.850082] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 716.850276] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 717.031204] env[61857]: INFO nova.compute.manager [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 9edf9e8c-5421-4018-afee-59e10cd335ef] Took 1.02 seconds to deallocate network for instance. [ 717.165386] env[61857]: DEBUG nova.compute.manager [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.335023] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.335356] env[61857]: DEBUG nova.compute.manager [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 717.338090] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.243s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.353244] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.687214] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.842629] env[61857]: DEBUG nova.compute.utils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.844668] env[61857]: DEBUG nova.compute.manager [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 717.844876] env[61857]: DEBUG nova.network.neutron [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 717.910351] env[61857]: DEBUG nova.policy [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f54c82d58ae4adabefd4ebba5d19263', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6be7236fbc844deb8011502c4910401f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 718.061496] env[61857]: INFO nova.scheduler.client.report [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Deleted allocations for instance 9edf9e8c-5421-4018-afee-59e10cd335ef [ 718.230617] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.230886] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.239342] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe611f9c-4a55-4e51-967f-9244bd192f9e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.248103] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67040f47-d8ba-4a46-a709-ef3f2521831a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.279834] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5298397-a30b-41f2-a561-12eed07bb538 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.284308] env[61857]: DEBUG nova.network.neutron [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Successfully created port: 2104d242-18d6-4e99-b6d0-1dd9efc2778b {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 718.289168] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5807410-b2d7-4197-b2fd-67c4fce23d9b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.303267] env[61857]: DEBUG nova.compute.provider_tree [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.345442] env[61857]: DEBUG nova.compute.manager [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 718.571726] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523ac9e8-6a17-4bca-b1f4-10689b6510d3 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "9edf9e8c-5421-4018-afee-59e10cd335ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.063s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.806650] env[61857]: DEBUG nova.scheduler.client.report [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.074715] env[61857]: DEBUG nova.compute.manager [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.315639] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.315639] env[61857]: ERROR nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b, please check neutron logs for more information. [ 719.315639] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] Traceback (most recent call last): [ 719.315639] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.315639] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] self.driver.spawn(context, instance, image_meta, [ 719.315639] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 719.315639] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.315639] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.315639] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] vm_ref = self.build_virtual_machine(instance, [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] for vif in network_info: [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] return self._sync_wrapper(fn, *args, **kwargs) [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] self.wait() [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] self[:] = self._gt.wait() [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] return self._exit_event.wait() [ 719.316106] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] current.throw(*self._exc) [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] result = function(*args, **kwargs) [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] return func(*args, **kwargs) [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] raise e [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] nwinfo = self.network_api.allocate_for_instance( [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] created_port_ids = self._update_ports_for_instance( [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 719.316459] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] with excutils.save_and_reraise_exception(): [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] self.force_reraise() [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] raise self.value [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] updated_port = self._update_port( [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] _ensure_no_port_binding_failure(port) [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] raise exception.PortBindingFailed(port_id=port['id']) [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] nova.exception.PortBindingFailed: Binding failed for port 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b, please check neutron logs for more information. [ 719.316806] env[61857]: ERROR nova.compute.manager [instance: e52972ca-877f-45b8-b79f-c834246a8857] [ 719.317134] env[61857]: DEBUG nova.compute.utils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Binding failed for port 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 719.320147] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.647s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.321024] env[61857]: DEBUG nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Build of instance e52972ca-877f-45b8-b79f-c834246a8857 was re-scheduled: Binding failed for port 0f9aef49-4bc9-4034-8e05-9e0eb76bdd2b, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 719.322237] env[61857]: DEBUG nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 719.323468] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Acquiring lock "refresh_cache-e52972ca-877f-45b8-b79f-c834246a8857" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.323741] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Acquired lock "refresh_cache-e52972ca-877f-45b8-b79f-c834246a8857" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.324038] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 719.357114] env[61857]: DEBUG nova.compute.manager [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 719.388199] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 719.388445] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 719.388602] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 719.388782] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 719.388928] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 719.389094] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 719.389307] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 719.389507] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 719.389627] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 719.389854] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 719.389952] env[61857]: DEBUG nova.virt.hardware [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 719.391122] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824bd066-1b4d-4f46-a5a2-1a0015551bc9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.399213] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014fb8d4-5bfc-45a1-83ff-6296897cae07 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.599567] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.844422] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.864014] env[61857]: DEBUG nova.compute.manager [req-88d4d227-a7db-46c4-bbff-b4bf0d2b6ca4 req-0a6b089e-ae40-4f5c-b8b4-7ad2a0c39499 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Received event network-vif-plugged-2104d242-18d6-4e99-b6d0-1dd9efc2778b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.864244] env[61857]: DEBUG oslo_concurrency.lockutils [req-88d4d227-a7db-46c4-bbff-b4bf0d2b6ca4 req-0a6b089e-ae40-4f5c-b8b4-7ad2a0c39499 service nova] Acquiring lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.864447] env[61857]: DEBUG oslo_concurrency.lockutils [req-88d4d227-a7db-46c4-bbff-b4bf0d2b6ca4 req-0a6b089e-ae40-4f5c-b8b4-7ad2a0c39499 service nova] Lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.864610] env[61857]: DEBUG oslo_concurrency.lockutils [req-88d4d227-a7db-46c4-bbff-b4bf0d2b6ca4 req-0a6b089e-ae40-4f5c-b8b4-7ad2a0c39499 service nova] Lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.864861] env[61857]: DEBUG nova.compute.manager [req-88d4d227-a7db-46c4-bbff-b4bf0d2b6ca4 req-0a6b089e-ae40-4f5c-b8b4-7ad2a0c39499 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] No waiting events found dispatching network-vif-plugged-2104d242-18d6-4e99-b6d0-1dd9efc2778b {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 719.864933] env[61857]: WARNING nova.compute.manager [req-88d4d227-a7db-46c4-bbff-b4bf0d2b6ca4 req-0a6b089e-ae40-4f5c-b8b4-7ad2a0c39499 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Received unexpected event network-vif-plugged-2104d242-18d6-4e99-b6d0-1dd9efc2778b for instance with vm_state building and task_state spawning. [ 719.944604] env[61857]: DEBUG nova.network.neutron [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Successfully updated port: 2104d242-18d6-4e99-b6d0-1dd9efc2778b {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 719.976147] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.220452] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8bed19-81dc-413b-8d34-16bb3e67f8fa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.229194] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e338798e-2dee-46fe-8cf2-d76a86db4736 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.262821] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d92aeb-33c3-478e-9b93-9a05f36a3e16 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.270060] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab33e57-8f06-41b5-9a59-34fc63103251 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.285132] env[61857]: DEBUG nova.compute.provider_tree [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.447679] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquiring lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.447925] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquired lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.448044] env[61857]: DEBUG nova.network.neutron [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 720.481690] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Releasing lock "refresh_cache-e52972ca-877f-45b8-b79f-c834246a8857" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.481940] env[61857]: DEBUG nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 720.482999] env[61857]: DEBUG nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.482999] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 720.495870] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.789482] env[61857]: DEBUG nova.scheduler.client.report [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.792356] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "16b447b5-426a-4478-9d44-ae32b41dee50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.792576] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.998090] env[61857]: DEBUG nova.network.neutron [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.000120] env[61857]: DEBUG nova.network.neutron [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.206287] env[61857]: DEBUG nova.network.neutron [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Updating instance_info_cache with network_info: [{"id": "2104d242-18d6-4e99-b6d0-1dd9efc2778b", "address": "fa:16:3e:95:af:9d", "network": {"id": "ca954f70-9b06-47d1-8347-d99e04452f69", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1065246308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6be7236fbc844deb8011502c4910401f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2104d242-18", "ovs_interfaceid": "2104d242-18d6-4e99-b6d0-1dd9efc2778b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.294026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.976s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.294648] env[61857]: ERROR nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 51b2e7bb-94cc-4ae3-a62a-e860d82899f9, please check neutron logs for more information. [ 721.294648] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Traceback (most recent call last): [ 721.294648] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.294648] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] self.driver.spawn(context, instance, image_meta, [ 721.294648] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 721.294648] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.294648] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.294648] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] vm_ref = self.build_virtual_machine(instance, [ 721.294648] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.294648] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.294648] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] for vif in network_info: [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] return self._sync_wrapper(fn, *args, **kwargs) [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] self.wait() [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] self[:] = self._gt.wait() [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] return self._exit_event.wait() [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] current.throw(*self._exc) [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.295039] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] result = function(*args, **kwargs) [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] return func(*args, **kwargs) [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] raise e [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] nwinfo = self.network_api.allocate_for_instance( [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] created_port_ids = self._update_ports_for_instance( [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] with excutils.save_and_reraise_exception(): [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] self.force_reraise() [ 721.295511] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.295933] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] raise self.value [ 721.295933] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 721.295933] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] updated_port = self._update_port( [ 721.295933] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.295933] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] _ensure_no_port_binding_failure(port) [ 721.295933] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.295933] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] raise exception.PortBindingFailed(port_id=port['id']) [ 721.295933] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] nova.exception.PortBindingFailed: Binding failed for port 51b2e7bb-94cc-4ae3-a62a-e860d82899f9, please check neutron logs for more information. [ 721.295933] env[61857]: ERROR nova.compute.manager [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] [ 721.295933] env[61857]: DEBUG nova.compute.utils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Binding failed for port 51b2e7bb-94cc-4ae3-a62a-e860d82899f9, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 721.296568] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.976s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.300745] env[61857]: DEBUG nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Build of instance 952d0a1c-3e94-42dc-9bc4-8063275afc82 was re-scheduled: Binding failed for port 51b2e7bb-94cc-4ae3-a62a-e860d82899f9, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 721.301177] env[61857]: DEBUG nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 721.301403] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Acquiring lock "refresh_cache-952d0a1c-3e94-42dc-9bc4-8063275afc82" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.301550] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Acquired lock "refresh_cache-952d0a1c-3e94-42dc-9bc4-8063275afc82" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.301710] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 721.502727] env[61857]: INFO nova.compute.manager [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] [instance: e52972ca-877f-45b8-b79f-c834246a8857] Took 1.02 seconds to deallocate network for instance. [ 721.709522] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Releasing lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.709800] env[61857]: DEBUG nova.compute.manager [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Instance network_info: |[{"id": "2104d242-18d6-4e99-b6d0-1dd9efc2778b", "address": "fa:16:3e:95:af:9d", "network": {"id": "ca954f70-9b06-47d1-8347-d99e04452f69", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1065246308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6be7236fbc844deb8011502c4910401f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2104d242-18", "ovs_interfaceid": "2104d242-18d6-4e99-b6d0-1dd9efc2778b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 721.710355] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:af:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cf5bfbae-a882-4d34-be33-b31e274b3077', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2104d242-18d6-4e99-b6d0-1dd9efc2778b', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 721.717964] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Creating folder: Project (6be7236fbc844deb8011502c4910401f). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 721.718279] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-70876ee6-f0e6-499d-8310-9ef10b475abb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.729530] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Created folder: Project (6be7236fbc844deb8011502c4910401f) in parent group-v214027. [ 721.729717] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Creating folder: Instances. Parent ref: group-v214044. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 721.729938] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44cb4039-9f78-4035-84fa-1ead0375602c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.738860] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Created folder: Instances in parent group-v214044. [ 721.739103] env[61857]: DEBUG oslo.service.loopingcall [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.739416] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 721.739488] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-470dfbfd-65c4-4e9e-b445-6d98b2cb52af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.757194] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 721.757194] env[61857]: value = "task-950724" [ 721.757194] env[61857]: _type = "Task" [ 721.757194] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.764545] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950724, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.825354] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.894259] env[61857]: DEBUG nova.compute.manager [req-47523ac3-6245-4869-914b-814a7eae57d7 req-74643f67-bd0b-4c81-a612-5b509eb49704 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Received event network-changed-2104d242-18d6-4e99-b6d0-1dd9efc2778b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 721.894481] env[61857]: DEBUG nova.compute.manager [req-47523ac3-6245-4869-914b-814a7eae57d7 req-74643f67-bd0b-4c81-a612-5b509eb49704 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Refreshing instance network info cache due to event network-changed-2104d242-18d6-4e99-b6d0-1dd9efc2778b. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 721.895105] env[61857]: DEBUG oslo_concurrency.lockutils [req-47523ac3-6245-4869-914b-814a7eae57d7 req-74643f67-bd0b-4c81-a612-5b509eb49704 service nova] Acquiring lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.895105] env[61857]: DEBUG oslo_concurrency.lockutils [req-47523ac3-6245-4869-914b-814a7eae57d7 req-74643f67-bd0b-4c81-a612-5b509eb49704 service nova] Acquired lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.895105] env[61857]: DEBUG nova.network.neutron [req-47523ac3-6245-4869-914b-814a7eae57d7 req-74643f67-bd0b-4c81-a612-5b509eb49704 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Refreshing network info cache for port 2104d242-18d6-4e99-b6d0-1dd9efc2778b {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 721.915112] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.171785] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04e7e04-1df6-4b4d-a245-6564eb46cc15 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.178956] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80ad1e3-35e1-41be-92db-4fd838a83cbb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.209224] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0aa1baa-0013-4242-a35e-3be5aef55ac0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.216089] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe9bf13-12ff-4971-9ac8-e07571f84d26 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.228775] env[61857]: DEBUG nova.compute.provider_tree [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.266187] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950724, 'name': CreateVM_Task, 'duration_secs': 0.279208} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.266359] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 722.272833] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.273008] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.273386] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 722.273633] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d6b0740-9024-4274-8756-b2d17c9b7c82 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.278086] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for the task: (returnval){ [ 722.278086] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52999d20-25bb-8db0-4bc5-35397371c835" [ 722.278086] env[61857]: _type = "Task" [ 722.278086] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.285017] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52999d20-25bb-8db0-4bc5-35397371c835, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.418846] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Releasing lock "refresh_cache-952d0a1c-3e94-42dc-9bc4-8063275afc82" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.419124] env[61857]: DEBUG nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 722.419315] env[61857]: DEBUG nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.419499] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 722.444064] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.532862] env[61857]: INFO nova.scheduler.client.report [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Deleted allocations for instance e52972ca-877f-45b8-b79f-c834246a8857 [ 722.664468] env[61857]: DEBUG nova.network.neutron [req-47523ac3-6245-4869-914b-814a7eae57d7 req-74643f67-bd0b-4c81-a612-5b509eb49704 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Updated VIF entry in instance network info cache for port 2104d242-18d6-4e99-b6d0-1dd9efc2778b. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 722.664737] env[61857]: DEBUG nova.network.neutron [req-47523ac3-6245-4869-914b-814a7eae57d7 req-74643f67-bd0b-4c81-a612-5b509eb49704 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Updating instance_info_cache with network_info: [{"id": "2104d242-18d6-4e99-b6d0-1dd9efc2778b", "address": "fa:16:3e:95:af:9d", "network": {"id": "ca954f70-9b06-47d1-8347-d99e04452f69", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1065246308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6be7236fbc844deb8011502c4910401f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2104d242-18", "ovs_interfaceid": "2104d242-18d6-4e99-b6d0-1dd9efc2778b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.731660] env[61857]: DEBUG nova.scheduler.client.report [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.788331] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52999d20-25bb-8db0-4bc5-35397371c835, 'name': SearchDatastore_Task, 'duration_secs': 0.010127} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.788627] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.788859] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 722.789094] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.789241] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.789421] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 722.789666] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d75cc7c9-3b36-4334-8c08-dcd4a1a02e55 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.797182] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 722.797345] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 722.798045] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a401169a-d714-46d0-9652-6f941e04fd60 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.803458] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for the task: (returnval){ [ 722.803458] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5200f5ac-1680-a3d9-147b-4e234c9e5ba5" [ 722.803458] env[61857]: _type = "Task" [ 722.803458] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.811154] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5200f5ac-1680-a3d9-147b-4e234c9e5ba5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.946956] env[61857]: DEBUG nova.network.neutron [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.043609] env[61857]: DEBUG oslo_concurrency.lockutils [None req-257ab1c6-f725-428e-995f-3d79d2827112 tempest-ServerActionsV293TestJSON-718660338 tempest-ServerActionsV293TestJSON-718660338-project-member] Lock "e52972ca-877f-45b8-b79f-c834246a8857" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.676s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.167551] env[61857]: DEBUG oslo_concurrency.lockutils [req-47523ac3-6245-4869-914b-814a7eae57d7 req-74643f67-bd0b-4c81-a612-5b509eb49704 service nova] Releasing lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.236411] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.940s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.237088] env[61857]: ERROR nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 99bdd524-9ab6-4c8a-8b03-dd16ea228712, please check neutron logs for more information. [ 723.237088] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Traceback (most recent call last): [ 723.237088] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.237088] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] self.driver.spawn(context, instance, image_meta, [ 723.237088] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 723.237088] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.237088] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.237088] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] vm_ref = self.build_virtual_machine(instance, [ 723.237088] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.237088] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.237088] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] for vif in network_info: [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] return self._sync_wrapper(fn, *args, **kwargs) [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] self.wait() [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] self[:] = self._gt.wait() [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] return self._exit_event.wait() [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] result = hub.switch() [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 723.237488] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] return self.greenlet.switch() [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] result = function(*args, **kwargs) [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] return func(*args, **kwargs) [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] raise e [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] nwinfo = self.network_api.allocate_for_instance( [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] created_port_ids = self._update_ports_for_instance( [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] with excutils.save_and_reraise_exception(): [ 723.237957] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] self.force_reraise() [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] raise self.value [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] updated_port = self._update_port( [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] _ensure_no_port_binding_failure(port) [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] raise exception.PortBindingFailed(port_id=port['id']) [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] nova.exception.PortBindingFailed: Binding failed for port 99bdd524-9ab6-4c8a-8b03-dd16ea228712, please check neutron logs for more information. [ 723.238415] env[61857]: ERROR nova.compute.manager [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] [ 723.238766] env[61857]: DEBUG nova.compute.utils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Binding failed for port 99bdd524-9ab6-4c8a-8b03-dd16ea228712, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 723.239031] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.110s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.241831] env[61857]: DEBUG nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Build of instance 768e0e7d-7ebe-4d7a-a971-4fb463731050 was re-scheduled: Binding failed for port 99bdd524-9ab6-4c8a-8b03-dd16ea228712, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 723.242263] env[61857]: DEBUG nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 723.242514] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Acquiring lock "refresh_cache-768e0e7d-7ebe-4d7a-a971-4fb463731050" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.242665] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Acquired lock "refresh_cache-768e0e7d-7ebe-4d7a-a971-4fb463731050" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.242823] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.313575] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5200f5ac-1680-a3d9-147b-4e234c9e5ba5, 'name': SearchDatastore_Task, 'duration_secs': 0.007984} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.314472] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c9f1e8c-99f7-4d84-9dd6-9316728ae8e2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.319523] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for the task: (returnval){ [ 723.319523] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]526ca4f9-b724-29b4-7b3c-1fa6a4240f80" [ 723.319523] env[61857]: _type = "Task" [ 723.319523] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.326073] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]526ca4f9-b724-29b4-7b3c-1fa6a4240f80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.449673] env[61857]: INFO nova.compute.manager [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] [instance: 952d0a1c-3e94-42dc-9bc4-8063275afc82] Took 1.03 seconds to deallocate network for instance. [ 723.546117] env[61857]: DEBUG nova.compute.manager [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.764526] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.835379] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]526ca4f9-b724-29b4-7b3c-1fa6a4240f80, 'name': SearchDatastore_Task, 'duration_secs': 0.008382} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.835655] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.835880] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 2c187f64-8a36-4dfd-94e3-8ea944dbac24/2c187f64-8a36-4dfd-94e3-8ea944dbac24.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 723.836155] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-246a97ab-4ae4-41b6-aa72-39c9b71be98a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.845451] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for the task: (returnval){ [ 723.845451] env[61857]: value = "task-950726" [ 723.845451] env[61857]: _type = "Task" [ 723.845451] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.856836] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950726, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.906822] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.067815] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.151903] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8380a2ba-e03c-43fa-95ed-fdb5c0dab1d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.164764] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e02802c3-d331-4005-8299-145d3edae046 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.198337] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38dd7a81-5207-4e68-a8df-5a88d65093f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.206555] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4afd18-8f8c-4a44-ba99-64c65d2ef781 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.223639] env[61857]: DEBUG nova.compute.provider_tree [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.355989] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950726, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460656} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.356277] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 2c187f64-8a36-4dfd-94e3-8ea944dbac24/2c187f64-8a36-4dfd-94e3-8ea944dbac24.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 724.356479] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 724.356726] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f9f9ce7a-d46a-46b6-be88-25ee4683ac8a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.362781] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for the task: (returnval){ [ 724.362781] env[61857]: value = "task-950727" [ 724.362781] env[61857]: _type = "Task" [ 724.362781] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.370862] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950727, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.410095] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Releasing lock "refresh_cache-768e0e7d-7ebe-4d7a-a971-4fb463731050" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.410095] env[61857]: DEBUG nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 724.410292] env[61857]: DEBUG nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.410438] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 724.426184] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.485879] env[61857]: INFO nova.scheduler.client.report [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Deleted allocations for instance 952d0a1c-3e94-42dc-9bc4-8063275afc82 [ 724.726598] env[61857]: DEBUG nova.scheduler.client.report [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.872625] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950727, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059124} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.872912] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 724.873711] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9f10a9-41c7-42d0-b045-907ebf5ded35 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.898030] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] 2c187f64-8a36-4dfd-94e3-8ea944dbac24/2c187f64-8a36-4dfd-94e3-8ea944dbac24.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 724.898324] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed8a3a69-bc9b-42b3-a1d1-408c1199c8d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.919327] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for the task: (returnval){ [ 724.919327] env[61857]: value = "task-950728" [ 724.919327] env[61857]: _type = "Task" [ 724.919327] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.927725] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950728, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.932068] env[61857]: DEBUG nova.network.neutron [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.994736] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fbfef2d-5efa-4746-aaaf-eceab0bfdaf5 tempest-ServersNegativeTestMultiTenantJSON-1689515525 tempest-ServersNegativeTestMultiTenantJSON-1689515525-project-member] Lock "952d0a1c-3e94-42dc-9bc4-8063275afc82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.055s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.231664] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.232405] env[61857]: ERROR nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c9b09025-f020-4647-9acb-5be89eb1ac86, please check neutron logs for more information. [ 725.232405] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Traceback (most recent call last): [ 725.232405] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.232405] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] self.driver.spawn(context, instance, image_meta, [ 725.232405] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 725.232405] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.232405] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.232405] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] vm_ref = self.build_virtual_machine(instance, [ 725.232405] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.232405] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.232405] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] for vif in network_info: [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] return self._sync_wrapper(fn, *args, **kwargs) [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] self.wait() [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] self[:] = self._gt.wait() [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] return self._exit_event.wait() [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] result = hub.switch() [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 725.232949] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] return self.greenlet.switch() [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] result = function(*args, **kwargs) [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] return func(*args, **kwargs) [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] raise e [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] nwinfo = self.network_api.allocate_for_instance( [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] created_port_ids = self._update_ports_for_instance( [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] with excutils.save_and_reraise_exception(): [ 725.233740] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] self.force_reraise() [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] raise self.value [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] updated_port = self._update_port( [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] _ensure_no_port_binding_failure(port) [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] raise exception.PortBindingFailed(port_id=port['id']) [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] nova.exception.PortBindingFailed: Binding failed for port c9b09025-f020-4647-9acb-5be89eb1ac86, please check neutron logs for more information. [ 725.234435] env[61857]: ERROR nova.compute.manager [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] [ 725.234931] env[61857]: DEBUG nova.compute.utils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Binding failed for port c9b09025-f020-4647-9acb-5be89eb1ac86, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 725.234931] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.978s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.236086] env[61857]: INFO nova.compute.claims [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.238916] env[61857]: DEBUG nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Build of instance b3fd3980-efe4-49a0-b656-32f1543e657c was re-scheduled: Binding failed for port c9b09025-f020-4647-9acb-5be89eb1ac86, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 725.239393] env[61857]: DEBUG nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 725.239648] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "refresh_cache-b3fd3980-efe4-49a0-b656-32f1543e657c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.239763] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquired lock "refresh_cache-b3fd3980-efe4-49a0-b656-32f1543e657c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.239921] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 725.429452] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950728, 'name': ReconfigVM_Task, 'duration_secs': 0.265837} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.429722] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Reconfigured VM instance instance-00000029 to attach disk [datastore2] 2c187f64-8a36-4dfd-94e3-8ea944dbac24/2c187f64-8a36-4dfd-94e3-8ea944dbac24.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 725.430385] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-904f26a7-719e-4c06-84fd-07c3eb24db22 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.432134] env[61857]: INFO nova.compute.manager [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] [instance: 768e0e7d-7ebe-4d7a-a971-4fb463731050] Took 1.02 seconds to deallocate network for instance. [ 725.439096] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for the task: (returnval){ [ 725.439096] env[61857]: value = "task-950729" [ 725.439096] env[61857]: _type = "Task" [ 725.439096] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.446809] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950729, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.497067] env[61857]: DEBUG nova.compute.manager [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.765515] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.931688] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.950199] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950729, 'name': Rename_Task, 'duration_secs': 0.150248} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.950628] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 725.950975] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3be3dc2-31e4-47b2-9df6-6a2c8d8de763 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.961028] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for the task: (returnval){ [ 725.961028] env[61857]: value = "task-950730" [ 725.961028] env[61857]: _type = "Task" [ 725.961028] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.970135] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950730, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.026395] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.433254] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Releasing lock "refresh_cache-b3fd3980-efe4-49a0-b656-32f1543e657c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.433415] env[61857]: DEBUG nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 726.433605] env[61857]: DEBUG nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.434153] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 726.467509] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.469430] env[61857]: INFO nova.scheduler.client.report [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Deleted allocations for instance 768e0e7d-7ebe-4d7a-a971-4fb463731050 [ 726.478207] env[61857]: DEBUG oslo_vmware.api [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950730, 'name': PowerOnVM_Task, 'duration_secs': 0.443742} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.481635] env[61857]: DEBUG nova.network.neutron [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.483411] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 726.483625] env[61857]: INFO nova.compute.manager [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Took 7.13 seconds to spawn the instance on the hypervisor. [ 726.483802] env[61857]: DEBUG nova.compute.manager [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 726.485277] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9672efd-1983-46c2-beca-6ba631f3c6f6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.718922] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992e74bd-e61e-4b23-ab27-1bc0cd703824 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.729920] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bb953c-5a3f-4e3e-9a36-63526f866a91 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.765910] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a670b53a-ff21-43cb-b55f-aad92177156e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.777027] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb891f1-8032-4f11-933c-63df1840e065 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.795685] env[61857]: DEBUG nova.compute.provider_tree [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.985634] env[61857]: INFO nova.compute.manager [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: b3fd3980-efe4-49a0-b656-32f1543e657c] Took 0.55 seconds to deallocate network for instance. [ 726.988574] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a44cee77-03cd-483d-989d-6e002e53e524 tempest-ServersTestFqdnHostnames-2005032792 tempest-ServersTestFqdnHostnames-2005032792-project-member] Lock "768e0e7d-7ebe-4d7a-a971-4fb463731050" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.380s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.015081] env[61857]: INFO nova.compute.manager [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Took 25.39 seconds to build instance. [ 727.300090] env[61857]: DEBUG nova.scheduler.client.report [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.496836] env[61857]: DEBUG nova.compute.manager [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 727.517904] env[61857]: DEBUG oslo_concurrency.lockutils [None req-820cfeab-a120-4997-b094-cb03c9a7b98a tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.602s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.662596] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "ece27044-3123-4685-a4ab-30eeb10337d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.662886] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "ece27044-3123-4685-a4ab-30eeb10337d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.810177] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.810719] env[61857]: DEBUG nova.compute.manager [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.813253] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.653s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.024697] env[61857]: DEBUG nova.compute.manager [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.033050] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.038148] env[61857]: INFO nova.scheduler.client.report [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Deleted allocations for instance b3fd3980-efe4-49a0-b656-32f1543e657c [ 728.297723] env[61857]: DEBUG nova.compute.manager [req-4162c811-eef4-4e78-9527-e5127128f1df req-9e758247-bbcf-41e3-9225-58fbafa4b922 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Received event network-changed-2104d242-18d6-4e99-b6d0-1dd9efc2778b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 728.297932] env[61857]: DEBUG nova.compute.manager [req-4162c811-eef4-4e78-9527-e5127128f1df req-9e758247-bbcf-41e3-9225-58fbafa4b922 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Refreshing instance network info cache due to event network-changed-2104d242-18d6-4e99-b6d0-1dd9efc2778b. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 728.298166] env[61857]: DEBUG oslo_concurrency.lockutils [req-4162c811-eef4-4e78-9527-e5127128f1df req-9e758247-bbcf-41e3-9225-58fbafa4b922 service nova] Acquiring lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.298304] env[61857]: DEBUG oslo_concurrency.lockutils [req-4162c811-eef4-4e78-9527-e5127128f1df req-9e758247-bbcf-41e3-9225-58fbafa4b922 service nova] Acquired lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.298466] env[61857]: DEBUG nova.network.neutron [req-4162c811-eef4-4e78-9527-e5127128f1df req-9e758247-bbcf-41e3-9225-58fbafa4b922 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Refreshing network info cache for port 2104d242-18d6-4e99-b6d0-1dd9efc2778b {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 728.318287] env[61857]: DEBUG nova.compute.utils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.323217] env[61857]: DEBUG nova.compute.manager [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Not allocating networking since 'none' was specified. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 728.550164] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8a50c5ea-ddea-4c14-8677-357ae94dd46e tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "b3fd3980-efe4-49a0-b656-32f1543e657c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.548s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.556151] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.734780] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa61d388-e777-4069-9f7c-46505f0edcac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.742610] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb24310a-167d-49b6-a79f-26e52a24151a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.773973] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd3a154-8df2-4f72-baf7-5356882ac098 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.781492] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75aa6a7a-704f-4954-be21-5cc9f6c2a167 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.794928] env[61857]: DEBUG nova.compute.provider_tree [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.823585] env[61857]: DEBUG nova.compute.manager [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 729.058588] env[61857]: DEBUG nova.compute.manager [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 729.156631] env[61857]: DEBUG nova.network.neutron [req-4162c811-eef4-4e78-9527-e5127128f1df req-9e758247-bbcf-41e3-9225-58fbafa4b922 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Updated VIF entry in instance network info cache for port 2104d242-18d6-4e99-b6d0-1dd9efc2778b. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 729.157112] env[61857]: DEBUG nova.network.neutron [req-4162c811-eef4-4e78-9527-e5127128f1df req-9e758247-bbcf-41e3-9225-58fbafa4b922 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Updating instance_info_cache with network_info: [{"id": "2104d242-18d6-4e99-b6d0-1dd9efc2778b", "address": "fa:16:3e:95:af:9d", "network": {"id": "ca954f70-9b06-47d1-8347-d99e04452f69", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1065246308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6be7236fbc844deb8011502c4910401f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2104d242-18", "ovs_interfaceid": "2104d242-18d6-4e99-b6d0-1dd9efc2778b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.298031] env[61857]: DEBUG nova.scheduler.client.report [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.589620] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.666077] env[61857]: DEBUG oslo_concurrency.lockutils [req-4162c811-eef4-4e78-9527-e5127128f1df req-9e758247-bbcf-41e3-9225-58fbafa4b922 service nova] Releasing lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.803416] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.990s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.804167] env[61857]: ERROR nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 619c90fa-1887-4f9c-b8a7-ad58472dbbf5, please check neutron logs for more information. [ 729.804167] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Traceback (most recent call last): [ 729.804167] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.804167] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] self.driver.spawn(context, instance, image_meta, [ 729.804167] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 729.804167] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.804167] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.804167] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] vm_ref = self.build_virtual_machine(instance, [ 729.804167] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.804167] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.804167] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] for vif in network_info: [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] return self._sync_wrapper(fn, *args, **kwargs) [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] self.wait() [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] self[:] = self._gt.wait() [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] return self._exit_event.wait() [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] current.throw(*self._exc) [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.804506] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] result = function(*args, **kwargs) [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] return func(*args, **kwargs) [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] raise e [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] nwinfo = self.network_api.allocate_for_instance( [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] created_port_ids = self._update_ports_for_instance( [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] with excutils.save_and_reraise_exception(): [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] self.force_reraise() [ 729.804854] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.805218] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] raise self.value [ 729.805218] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 729.805218] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] updated_port = self._update_port( [ 729.805218] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.805218] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] _ensure_no_port_binding_failure(port) [ 729.805218] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.805218] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] raise exception.PortBindingFailed(port_id=port['id']) [ 729.805218] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] nova.exception.PortBindingFailed: Binding failed for port 619c90fa-1887-4f9c-b8a7-ad58472dbbf5, please check neutron logs for more information. [ 729.805218] env[61857]: ERROR nova.compute.manager [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] [ 729.805218] env[61857]: DEBUG nova.compute.utils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Binding failed for port 619c90fa-1887-4f9c-b8a7-ad58472dbbf5, please check neutron logs for more information. {{(pid=61857) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 729.807193] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.562s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.808843] env[61857]: INFO nova.compute.claims [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.815582] env[61857]: DEBUG nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Build of instance e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd was re-scheduled: Binding failed for port 619c90fa-1887-4f9c-b8a7-ad58472dbbf5, please check neutron logs for more information. {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 729.816227] env[61857]: DEBUG nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Unplugging VIFs for instance {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 729.820023] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Acquiring lock "refresh_cache-e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.820023] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Acquired lock "refresh_cache-e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.820023] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 729.837183] env[61857]: DEBUG nova.compute.manager [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.875178] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.875441] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.875736] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.875877] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.876407] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.876705] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.877051] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.877377] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.878329] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.878632] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.879093] env[61857]: DEBUG nova.virt.hardware [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.881801] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857588e4-9c20-4d99-acfd-56d94f72e2df {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.893837] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ee7c2c-d598-4ab9-92a3-1a580935eb7d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.919514] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 729.930343] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Creating folder: Project (ac4fbd1eb8484f7b8e5269c394a9c043). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 729.930931] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a982e11-eef1-43a4-9d70-360db0156d3d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.944203] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Created folder: Project (ac4fbd1eb8484f7b8e5269c394a9c043) in parent group-v214027. [ 729.944203] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Creating folder: Instances. Parent ref: group-v214047. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 729.944203] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-39ab683e-2da3-4a52-a5b0-6c1ccb530601 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.958028] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Created folder: Instances in parent group-v214047. [ 729.958028] env[61857]: DEBUG oslo.service.loopingcall [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.958028] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 729.958838] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68f41cd5-a7b5-43df-80b2-c95d45c02882 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.980732] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 729.980732] env[61857]: value = "task-950733" [ 729.980732] env[61857]: _type = "Task" [ 729.980732] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.988567] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950733, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.362024] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.496106] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950733, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.583225] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.996712] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950733, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.088999] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Releasing lock "refresh_cache-e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.089211] env[61857]: DEBUG nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61857) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 731.089653] env[61857]: DEBUG nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.089653] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 731.127240] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.277652] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b35352-267f-4712-a293-e1ca35d0c7ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.285556] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089e5770-c6c2-45f8-b5b2-c96d7c9b87d3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.320382] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b267555-1c74-4520-9719-5defa372a896 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.332187] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3acc07-7759-45ca-a297-7d86f3c1b4f0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.345506] env[61857]: DEBUG nova.compute.provider_tree [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.493177] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950733, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.629550] env[61857]: DEBUG nova.network.neutron [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.852869] env[61857]: DEBUG nova.scheduler.client.report [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.994105] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950733, 'name': CreateVM_Task, 'duration_secs': 1.84968} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.994105] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 731.994105] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.994377] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.994464] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 731.994770] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc30974b-b134-4548-9be0-4bd7e5f696bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.001515] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for the task: (returnval){ [ 732.001515] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523fba0e-476c-5604-a43c-7af7ff179e74" [ 732.001515] env[61857]: _type = "Task" [ 732.001515] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.010763] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523fba0e-476c-5604-a43c-7af7ff179e74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.133518] env[61857]: INFO nova.compute.manager [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] [instance: e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd] Took 1.04 seconds to deallocate network for instance. [ 732.359892] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.360535] env[61857]: DEBUG nova.compute.manager [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 732.364792] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.012s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.364956] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.365118] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 732.365443] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.678s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.366858] env[61857]: INFO nova.compute.claims [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.370178] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad10ebf3-b7f7-4e84-a1aa-8a6fac540d91 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.380700] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af1b334-8d2a-447e-813b-372353809739 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.397034] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb3563d-90c3-4f2c-ac19-5ec30d1d7528 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.404868] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e83060-ac78-4d2d-917f-aef653fe311a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.436691] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181639MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 732.436806] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.511081] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523fba0e-476c-5604-a43c-7af7ff179e74, 'name': SearchDatastore_Task, 'duration_secs': 0.011393} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.511418] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.511687] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 732.511923] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.512102] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.512320] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 732.512588] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99077d9e-bf5b-4ba4-9b20-de30607f5184 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.519954] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 732.520180] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 732.521411] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed588e17-c759-446a-b9c6-497243109675 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.526241] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for the task: (returnval){ [ 732.526241] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ef9a82-2c07-6739-989b-213800375c8b" [ 732.526241] env[61857]: _type = "Task" [ 732.526241] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.535083] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ef9a82-2c07-6739-989b-213800375c8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.681311] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.681581] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.871730] env[61857]: DEBUG nova.compute.utils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 732.873699] env[61857]: DEBUG nova.compute.manager [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 732.873926] env[61857]: DEBUG nova.network.neutron [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 732.963948] env[61857]: DEBUG nova.policy [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e3e189b9b0946bfa6fab7e0680ffd79', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e0848df81fc84d25a4bb79d6f0b4b7cb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 733.039626] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ef9a82-2c07-6739-989b-213800375c8b, 'name': SearchDatastore_Task, 'duration_secs': 0.008501} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.042379] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edb0df5e-e661-4c19-9f12-50a255511fbd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.047979] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for the task: (returnval){ [ 733.047979] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5243bd66-831f-7e9f-406b-8445ec992dbf" [ 733.047979] env[61857]: _type = "Task" [ 733.047979] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.056507] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5243bd66-831f-7e9f-406b-8445ec992dbf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.179130] env[61857]: INFO nova.scheduler.client.report [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Deleted allocations for instance e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd [ 733.377122] env[61857]: DEBUG nova.compute.manager [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 733.415375] env[61857]: DEBUG nova.network.neutron [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Successfully created port: e656ed86-dc81-4383-8dcd-254c4a462021 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 733.565105] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5243bd66-831f-7e9f-406b-8445ec992dbf, 'name': SearchDatastore_Task, 'duration_secs': 0.009501} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.565376] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.565641] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 6304714b-8988-4798-990f-d3e19774eb8f/6304714b-8988-4798-990f-d3e19774eb8f.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 733.565903] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32964067-192f-495b-b302-3b6fdcc8c6da {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.579050] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for the task: (returnval){ [ 733.579050] env[61857]: value = "task-950734" [ 733.579050] env[61857]: _type = "Task" [ 733.579050] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.592215] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950734, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.689857] env[61857]: DEBUG oslo_concurrency.lockutils [None req-79f7ec35-efc8-4a04-96c3-f636894ab0cd tempest-InstanceActionsV221TestJSON-714178894 tempest-InstanceActionsV221TestJSON-714178894-project-member] Lock "e7bae1d1-160f-4ea0-b5e6-1ac31f3c91bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.667s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.848699] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ff0684-4cfd-4883-85ff-bea2e9463949 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.857030] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa57e0f-cbfa-4cd7-9845-12de18cd1594 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.900599] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6250e8cb-a49b-4ddf-94f0-989be8b24b8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.910148] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68215c02-96da-4740-b61c-50e79b17a233 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.932417] env[61857]: DEBUG nova.compute.provider_tree [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.088020] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950734, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.411935} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.088363] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 6304714b-8988-4798-990f-d3e19774eb8f/6304714b-8988-4798-990f-d3e19774eb8f.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 734.088537] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.089118] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db43d320-c77f-4f52-8219-ac3d9b72e69f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.094915] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for the task: (returnval){ [ 734.094915] env[61857]: value = "task-950735" [ 734.094915] env[61857]: _type = "Task" [ 734.094915] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.103553] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950735, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.195298] env[61857]: DEBUG nova.compute.manager [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.405850] env[61857]: DEBUG nova.compute.manager [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 734.435677] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.438395] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.438395] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.438395] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.438395] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.438395] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.438395] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.438691] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.438691] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.438691] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.438691] env[61857]: DEBUG nova.virt.hardware [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.438819] env[61857]: DEBUG nova.scheduler.client.report [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.444181] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73f99a5-086a-46de-b91f-1b9bacf84e78 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.456785] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5315f484-3469-472c-9c49-0c87728ffc17 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.606337] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950735, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063868} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.606564] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 734.607350] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e367442a-dfa9-4322-9a6d-a6f56198bfa0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.629608] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 6304714b-8988-4798-990f-d3e19774eb8f/6304714b-8988-4798-990f-d3e19774eb8f.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 734.629970] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ae8ea84-19f0-4b02-aae4-2171a71fbe23 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.651641] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for the task: (returnval){ [ 734.651641] env[61857]: value = "task-950736" [ 734.651641] env[61857]: _type = "Task" [ 734.651641] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.660366] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950736, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.726958] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.951019] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.583s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.951019] env[61857]: DEBUG nova.compute.manager [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 734.952412] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.353s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.954054] env[61857]: INFO nova.compute.claims [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.161845] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950736, 'name': ReconfigVM_Task, 'duration_secs': 0.272206} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.162166] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 6304714b-8988-4798-990f-d3e19774eb8f/6304714b-8988-4798-990f-d3e19774eb8f.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 735.162778] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-22737d80-15db-425d-969c-0297dbf6b664 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.169782] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for the task: (returnval){ [ 735.169782] env[61857]: value = "task-950737" [ 735.169782] env[61857]: _type = "Task" [ 735.169782] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.177463] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950737, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.427403] env[61857]: DEBUG nova.compute.manager [req-d08934e9-aad2-4023-93b4-424ebd63cb43 req-910a3c7d-d13f-401e-b45a-0d76146f4d49 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Received event network-vif-plugged-e656ed86-dc81-4383-8dcd-254c4a462021 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 735.427403] env[61857]: DEBUG oslo_concurrency.lockutils [req-d08934e9-aad2-4023-93b4-424ebd63cb43 req-910a3c7d-d13f-401e-b45a-0d76146f4d49 service nova] Acquiring lock "11adaee5-8e56-4679-8373-5d5690a44ca9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.427403] env[61857]: DEBUG oslo_concurrency.lockutils [req-d08934e9-aad2-4023-93b4-424ebd63cb43 req-910a3c7d-d13f-401e-b45a-0d76146f4d49 service nova] Lock "11adaee5-8e56-4679-8373-5d5690a44ca9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.427403] env[61857]: DEBUG oslo_concurrency.lockutils [req-d08934e9-aad2-4023-93b4-424ebd63cb43 req-910a3c7d-d13f-401e-b45a-0d76146f4d49 service nova] Lock "11adaee5-8e56-4679-8373-5d5690a44ca9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.427403] env[61857]: DEBUG nova.compute.manager [req-d08934e9-aad2-4023-93b4-424ebd63cb43 req-910a3c7d-d13f-401e-b45a-0d76146f4d49 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] No waiting events found dispatching network-vif-plugged-e656ed86-dc81-4383-8dcd-254c4a462021 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 735.427608] env[61857]: WARNING nova.compute.manager [req-d08934e9-aad2-4023-93b4-424ebd63cb43 req-910a3c7d-d13f-401e-b45a-0d76146f4d49 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Received unexpected event network-vif-plugged-e656ed86-dc81-4383-8dcd-254c4a462021 for instance with vm_state building and task_state spawning. [ 735.463800] env[61857]: DEBUG nova.compute.utils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.472295] env[61857]: DEBUG nova.compute.manager [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 735.472295] env[61857]: DEBUG nova.network.neutron [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 735.511397] env[61857]: DEBUG nova.network.neutron [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Successfully updated port: e656ed86-dc81-4383-8dcd-254c4a462021 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 735.582471] env[61857]: DEBUG nova.policy [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7aafb3356c5a48dcb36bf06d45851c3e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e3643abae2e40a99059603131e0b7b1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 735.685615] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950737, 'name': Rename_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.976188] env[61857]: DEBUG nova.compute.manager [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.014805] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquiring lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.015029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquired lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.015979] env[61857]: DEBUG nova.network.neutron [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.102359] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.103025] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.134603] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "ddd4c436-405f-49f7-8c9b-de3b71725f63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.135046] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "ddd4c436-405f-49f7-8c9b-de3b71725f63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.183767] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950737, 'name': Rename_Task, 'duration_secs': 0.844803} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.184582] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 736.184582] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67fa96d5-2065-40db-93ca-1fd1bf7d0300 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.191557] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for the task: (returnval){ [ 736.191557] env[61857]: value = "task-950738" [ 736.191557] env[61857]: _type = "Task" [ 736.191557] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.201753] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "7a9252da-2584-40ed-9d28-ca7341ed5165" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.201977] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "7a9252da-2584-40ed-9d28-ca7341ed5165" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.212323] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950738, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.233260] env[61857]: DEBUG nova.network.neutron [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Successfully created port: 03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.572264] env[61857]: DEBUG nova.network.neutron [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.589565] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1aa74c-fa13-4b43-9af1-b08013ba59c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.599109] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae150cd3-fa30-43c1-a895-bc3c2a4a7874 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.633337] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c705e295-33ae-4673-bc30-31ca0a20cea8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.641060] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804bd70f-b847-4d14-97a0-d044a9e78b4d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.655388] env[61857]: DEBUG nova.compute.provider_tree [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.704504] env[61857]: DEBUG oslo_vmware.api [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950738, 'name': PowerOnVM_Task, 'duration_secs': 0.411328} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.704769] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 736.704966] env[61857]: INFO nova.compute.manager [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Took 6.87 seconds to spawn the instance on the hypervisor. [ 736.705161] env[61857]: DEBUG nova.compute.manager [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 736.705914] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a25c24d-ba4f-4795-aa5f-8ae5dddb7b9a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.836586] env[61857]: DEBUG nova.network.neutron [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Updating instance_info_cache with network_info: [{"id": "e656ed86-dc81-4383-8dcd-254c4a462021", "address": "fa:16:3e:14:73:a6", "network": {"id": "32c36dba-b78e-4e41-ae8c-40f4b673555f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1785406565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e0848df81fc84d25a4bb79d6f0b4b7cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape656ed86-dc", "ovs_interfaceid": "e656ed86-dc81-4383-8dcd-254c4a462021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.989061] env[61857]: DEBUG nova.compute.manager [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.019175] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.019441] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.019599] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.019775] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.019918] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.020409] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.020664] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.020829] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.021012] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.021198] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.021390] env[61857]: DEBUG nova.virt.hardware [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.022236] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92c56ab-104a-446f-849a-a39513d49d56 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.030195] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34314568-54a5-4433-bfaa-9684ff39d317 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.163739] env[61857]: DEBUG nova.scheduler.client.report [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.222984] env[61857]: INFO nova.compute.manager [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Took 25.99 seconds to build instance. [ 737.340466] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Releasing lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.340808] env[61857]: DEBUG nova.compute.manager [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Instance network_info: |[{"id": "e656ed86-dc81-4383-8dcd-254c4a462021", "address": "fa:16:3e:14:73:a6", "network": {"id": "32c36dba-b78e-4e41-ae8c-40f4b673555f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1785406565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e0848df81fc84d25a4bb79d6f0b4b7cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape656ed86-dc", "ovs_interfaceid": "e656ed86-dc81-4383-8dcd-254c4a462021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 737.341268] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:73:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2a111da1-447a-4722-9662-b0d6a2886a65', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e656ed86-dc81-4383-8dcd-254c4a462021', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 737.349021] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Creating folder: Project (e0848df81fc84d25a4bb79d6f0b4b7cb). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 737.349483] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14258593-3379-4117-aea1-5eccc882d23a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.362300] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Created folder: Project (e0848df81fc84d25a4bb79d6f0b4b7cb) in parent group-v214027. [ 737.362524] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Creating folder: Instances. Parent ref: group-v214050. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 737.362760] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d33078f-8f1b-4fc1-bcee-d109237e6ba7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.373217] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Created folder: Instances in parent group-v214050. [ 737.373571] env[61857]: DEBUG oslo.service.loopingcall [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.373701] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 737.373892] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3709787-faaf-43f0-b900-1ee7d1f2bf82 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.392975] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.392975] env[61857]: value = "task-950741" [ 737.392975] env[61857]: _type = "Task" [ 737.392975] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.401296] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950741, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.458313] env[61857]: DEBUG nova.compute.manager [req-daab5e1e-b0af-4144-92fe-381c0dc937ed req-5b297f0d-4877-4445-b9fe-e0b6e6444725 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Received event network-changed-e656ed86-dc81-4383-8dcd-254c4a462021 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 737.458313] env[61857]: DEBUG nova.compute.manager [req-daab5e1e-b0af-4144-92fe-381c0dc937ed req-5b297f0d-4877-4445-b9fe-e0b6e6444725 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Refreshing instance network info cache due to event network-changed-e656ed86-dc81-4383-8dcd-254c4a462021. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 737.458659] env[61857]: DEBUG oslo_concurrency.lockutils [req-daab5e1e-b0af-4144-92fe-381c0dc937ed req-5b297f0d-4877-4445-b9fe-e0b6e6444725 service nova] Acquiring lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.458913] env[61857]: DEBUG oslo_concurrency.lockutils [req-daab5e1e-b0af-4144-92fe-381c0dc937ed req-5b297f0d-4877-4445-b9fe-e0b6e6444725 service nova] Acquired lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.459110] env[61857]: DEBUG nova.network.neutron [req-daab5e1e-b0af-4144-92fe-381c0dc937ed req-5b297f0d-4877-4445-b9fe-e0b6e6444725 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Refreshing network info cache for port e656ed86-dc81-4383-8dcd-254c4a462021 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 737.671818] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.719s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.672468] env[61857]: DEBUG nova.compute.manager [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 737.679560] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.609s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.679560] env[61857]: INFO nova.compute.claims [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.724794] env[61857]: DEBUG oslo_concurrency.lockutils [None req-83f4ba09-a985-4174-817e-fea022301c2d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lock "6304714b-8988-4798-990f-d3e19774eb8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.540s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.906885] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950741, 'name': CreateVM_Task, 'duration_secs': 0.346269} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.907764] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 737.907821] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.907986] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.908780] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 737.908780] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b29fdcb7-5094-4629-a218-170dafd23180 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.916520] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for the task: (returnval){ [ 737.916520] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52981972-5962-d775-29f7-5880e33af276" [ 737.916520] env[61857]: _type = "Task" [ 737.916520] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.929662] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52981972-5962-d775-29f7-5880e33af276, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.046592] env[61857]: DEBUG nova.network.neutron [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Successfully updated port: 03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 738.178107] env[61857]: DEBUG nova.compute.utils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.179404] env[61857]: DEBUG nova.compute.manager [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.179573] env[61857]: DEBUG nova.network.neutron [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 738.220998] env[61857]: DEBUG nova.compute.manager [req-23ac9916-c546-4d13-8d68-39c4547ec944 req-9941b17f-3004-434c-bcec-e2ad108cee4c service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Received event network-vif-plugged-03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 738.221239] env[61857]: DEBUG oslo_concurrency.lockutils [req-23ac9916-c546-4d13-8d68-39c4547ec944 req-9941b17f-3004-434c-bcec-e2ad108cee4c service nova] Acquiring lock "e88aed62-6561-4e29-b041-46b19cc5ef63-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.221447] env[61857]: DEBUG oslo_concurrency.lockutils [req-23ac9916-c546-4d13-8d68-39c4547ec944 req-9941b17f-3004-434c-bcec-e2ad108cee4c service nova] Lock "e88aed62-6561-4e29-b041-46b19cc5ef63-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.221605] env[61857]: DEBUG oslo_concurrency.lockutils [req-23ac9916-c546-4d13-8d68-39c4547ec944 req-9941b17f-3004-434c-bcec-e2ad108cee4c service nova] Lock "e88aed62-6561-4e29-b041-46b19cc5ef63-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.221767] env[61857]: DEBUG nova.compute.manager [req-23ac9916-c546-4d13-8d68-39c4547ec944 req-9941b17f-3004-434c-bcec-e2ad108cee4c service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] No waiting events found dispatching network-vif-plugged-03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 738.221922] env[61857]: WARNING nova.compute.manager [req-23ac9916-c546-4d13-8d68-39c4547ec944 req-9941b17f-3004-434c-bcec-e2ad108cee4c service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Received unexpected event network-vif-plugged-03542c83-4268-49c5-a829-9d7750f2a70a for instance with vm_state building and task_state spawning. [ 738.227188] env[61857]: DEBUG nova.compute.manager [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.232170] env[61857]: DEBUG nova.policy [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd169a99b1b904c589b4a062de88b5c6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a845df4e867940639f4ff7b6a092e301', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.371816] env[61857]: DEBUG nova.network.neutron [req-daab5e1e-b0af-4144-92fe-381c0dc937ed req-5b297f0d-4877-4445-b9fe-e0b6e6444725 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Updated VIF entry in instance network info cache for port e656ed86-dc81-4383-8dcd-254c4a462021. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 738.371816] env[61857]: DEBUG nova.network.neutron [req-daab5e1e-b0af-4144-92fe-381c0dc937ed req-5b297f0d-4877-4445-b9fe-e0b6e6444725 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Updating instance_info_cache with network_info: [{"id": "e656ed86-dc81-4383-8dcd-254c4a462021", "address": "fa:16:3e:14:73:a6", "network": {"id": "32c36dba-b78e-4e41-ae8c-40f4b673555f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1785406565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e0848df81fc84d25a4bb79d6f0b4b7cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape656ed86-dc", "ovs_interfaceid": "e656ed86-dc81-4383-8dcd-254c4a462021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.427571] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52981972-5962-d775-29f7-5880e33af276, 'name': SearchDatastore_Task, 'duration_secs': 0.009367} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.427864] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.428101] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 738.428331] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.428474] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.428645] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.429125] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2225bf6b-3c24-4d1a-81ee-ba9be78a1005 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.437604] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.437944] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 738.438574] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adde687e-ec26-4427-abe0-9b578f5ab2ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.444258] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for the task: (returnval){ [ 738.444258] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5296e0b7-3a9b-6e0e-2e08-603d8dfef504" [ 738.444258] env[61857]: _type = "Task" [ 738.444258] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.451898] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5296e0b7-3a9b-6e0e-2e08-603d8dfef504, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.553149] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.553149] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquired lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.553149] env[61857]: DEBUG nova.network.neutron [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.574595] env[61857]: DEBUG nova.network.neutron [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Successfully created port: 289c5ef4-b619-4720-bfc8-58a7b733beee {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 738.645019] env[61857]: DEBUG nova.compute.manager [None req-301d0c33-bcf1-416c-83b8-551a2730104d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 738.645019] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e26a768b-d44f-49ce-861b-36164ef894e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.686022] env[61857]: DEBUG nova.compute.manager [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 738.761305] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.873677] env[61857]: DEBUG oslo_concurrency.lockutils [req-daab5e1e-b0af-4144-92fe-381c0dc937ed req-5b297f0d-4877-4445-b9fe-e0b6e6444725 service nova] Releasing lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.966479] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5296e0b7-3a9b-6e0e-2e08-603d8dfef504, 'name': SearchDatastore_Task, 'duration_secs': 0.007979} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.967315] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a78dfd3-a843-4c21-8776-701afbb7e7cd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.974616] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for the task: (returnval){ [ 738.974616] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525dbd49-4c32-9742-e0ec-6806c1f49132" [ 738.974616] env[61857]: _type = "Task" [ 738.974616] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.982822] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquiring lock "6304714b-8988-4798-990f-d3e19774eb8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.983672] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lock "6304714b-8988-4798-990f-d3e19774eb8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.983672] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquiring lock "6304714b-8988-4798-990f-d3e19774eb8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.983672] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lock "6304714b-8988-4798-990f-d3e19774eb8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.983672] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lock "6304714b-8988-4798-990f-d3e19774eb8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.985300] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525dbd49-4c32-9742-e0ec-6806c1f49132, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.985630] env[61857]: INFO nova.compute.manager [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Terminating instance [ 738.989555] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquiring lock "refresh_cache-6304714b-8988-4798-990f-d3e19774eb8f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.989729] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquired lock "refresh_cache-6304714b-8988-4798-990f-d3e19774eb8f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.989886] env[61857]: DEBUG nova.network.neutron [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 739.115620] env[61857]: DEBUG nova.network.neutron [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.164548] env[61857]: INFO nova.compute.manager [None req-301d0c33-bcf1-416c-83b8-551a2730104d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] instance snapshotting [ 739.164548] env[61857]: DEBUG nova.objects.instance [None req-301d0c33-bcf1-416c-83b8-551a2730104d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lazy-loading 'flavor' on Instance uuid 6304714b-8988-4798-990f-d3e19774eb8f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 739.246957] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb4041b-902a-4889-b05a-2b2e730594ea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.259023] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1219ed31-001f-4f90-9121-bbe973aff186 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.295029] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c607b3c7-a12e-4e96-93be-fb5d57164a4a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.302791] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526d647e-0cf5-41e3-8505-7898c211aa8c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.318886] env[61857]: DEBUG nova.compute.provider_tree [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.385260] env[61857]: DEBUG nova.network.neutron [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Updating instance_info_cache with network_info: [{"id": "03542c83-4268-49c5-a829-9d7750f2a70a", "address": "fa:16:3e:03:ac:df", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03542c83-42", "ovs_interfaceid": "03542c83-4268-49c5-a829-9d7750f2a70a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.487236] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525dbd49-4c32-9742-e0ec-6806c1f49132, 'name': SearchDatastore_Task, 'duration_secs': 0.008839} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.488435] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.488435] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 11adaee5-8e56-4679-8373-5d5690a44ca9/11adaee5-8e56-4679-8373-5d5690a44ca9.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 739.488435] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6569e648-420b-41f4-96fa-f455135d05a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.497060] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for the task: (returnval){ [ 739.497060] env[61857]: value = "task-950742" [ 739.497060] env[61857]: _type = "Task" [ 739.497060] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.506764] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950742, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.521295] env[61857]: DEBUG nova.network.neutron [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.610086] env[61857]: DEBUG nova.network.neutron [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.669897] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc907695-f091-4ad1-b465-49a9972c33a5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.689802] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3312ef5a-1494-496e-a398-561395c89ed8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.702908] env[61857]: DEBUG nova.compute.manager [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 739.746145] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.746486] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.746651] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.746835] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.747016] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.747326] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.747413] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.747572] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.747740] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.748523] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.748624] env[61857]: DEBUG nova.virt.hardware [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.750750] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5ccc57-6ebd-43bb-a8ac-188688025985 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.760366] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27e6c01-b0ac-49d0-b61d-9bd09474956d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.822582] env[61857]: DEBUG nova.scheduler.client.report [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.887512] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Releasing lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.887907] env[61857]: DEBUG nova.compute.manager [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Instance network_info: |[{"id": "03542c83-4268-49c5-a829-9d7750f2a70a", "address": "fa:16:3e:03:ac:df", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03542c83-42", "ovs_interfaceid": "03542c83-4268-49c5-a829-9d7750f2a70a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 739.888609] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:ac:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4df917f7-847a-4c0e-b0e3-69a52e4a1554', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03542c83-4268-49c5-a829-9d7750f2a70a', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 739.898073] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Creating folder: Project (0e3643abae2e40a99059603131e0b7b1). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 739.898512] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4ba1769-c13d-4963-90f8-67222b74acef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.912674] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Created folder: Project (0e3643abae2e40a99059603131e0b7b1) in parent group-v214027. [ 739.912674] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Creating folder: Instances. Parent ref: group-v214053. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 739.913131] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f5c73fe2-77ae-4389-9d88-3a8d816d1dd9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.923917] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Created folder: Instances in parent group-v214053. [ 739.924245] env[61857]: DEBUG oslo.service.loopingcall [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.924456] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 739.924672] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2166a4b-9bef-4bf8-85be-1309800ae298 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.946053] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 739.946053] env[61857]: value = "task-950745" [ 739.946053] env[61857]: _type = "Task" [ 739.946053] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.957638] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950745, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.011319] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950742, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488429} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.011642] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 11adaee5-8e56-4679-8373-5d5690a44ca9/11adaee5-8e56-4679-8373-5d5690a44ca9.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 740.011867] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 740.012136] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b3799c3-2f35-4334-8bda-813c7c37b713 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.018714] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for the task: (returnval){ [ 740.018714] env[61857]: value = "task-950746" [ 740.018714] env[61857]: _type = "Task" [ 740.018714] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.027198] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950746, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.113354] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Releasing lock "refresh_cache-6304714b-8988-4798-990f-d3e19774eb8f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.113750] env[61857]: DEBUG nova.compute.manager [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 740.113940] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 740.114859] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b7c259-5df3-4a60-9acf-a5dbc4ddcbe1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.122753] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 740.122753] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77cf3c19-fdee-4b85-b4df-d2769b27d2c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.129451] env[61857]: DEBUG oslo_vmware.api [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for the task: (returnval){ [ 740.129451] env[61857]: value = "task-950747" [ 740.129451] env[61857]: _type = "Task" [ 740.129451] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.137951] env[61857]: DEBUG oslo_vmware.api [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950747, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.203495] env[61857]: DEBUG nova.compute.manager [None req-301d0c33-bcf1-416c-83b8-551a2730104d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Instance disappeared during snapshot {{(pid=61857) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 740.329243] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.652s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.329243] env[61857]: DEBUG nova.compute.manager [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 740.336168] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.307s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.336168] env[61857]: INFO nova.compute.claims [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.390389] env[61857]: DEBUG nova.compute.manager [None req-301d0c33-bcf1-416c-83b8-551a2730104d tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Found 0 images (rotation: 2) {{(pid=61857) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 740.456464] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950745, 'name': CreateVM_Task, 'duration_secs': 0.430933} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.456658] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 740.457318] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.457480] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.457802] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 740.458084] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e85ae1d6-e762-43a8-984f-d38db69c5a68 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.462709] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 740.462709] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5285af19-1f2b-3e9f-d03a-7b2b7e4b88e7" [ 740.462709] env[61857]: _type = "Task" [ 740.462709] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.472404] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5285af19-1f2b-3e9f-d03a-7b2b7e4b88e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.503364] env[61857]: DEBUG nova.compute.manager [req-528082fe-dcf3-4023-8852-23efe34438fc req-f11dcbd6-6f98-44d4-a316-850d1274aaf5 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Received event network-changed-03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 740.503364] env[61857]: DEBUG nova.compute.manager [req-528082fe-dcf3-4023-8852-23efe34438fc req-f11dcbd6-6f98-44d4-a316-850d1274aaf5 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Refreshing instance network info cache due to event network-changed-03542c83-4268-49c5-a829-9d7750f2a70a. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 740.503364] env[61857]: DEBUG oslo_concurrency.lockutils [req-528082fe-dcf3-4023-8852-23efe34438fc req-f11dcbd6-6f98-44d4-a316-850d1274aaf5 service nova] Acquiring lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.503364] env[61857]: DEBUG oslo_concurrency.lockutils [req-528082fe-dcf3-4023-8852-23efe34438fc req-f11dcbd6-6f98-44d4-a316-850d1274aaf5 service nova] Acquired lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.504391] env[61857]: DEBUG nova.network.neutron [req-528082fe-dcf3-4023-8852-23efe34438fc req-f11dcbd6-6f98-44d4-a316-850d1274aaf5 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Refreshing network info cache for port 03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 740.530051] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950746, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074835} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.530381] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 740.531242] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a448a7d-601e-4708-9ee0-1c68658bdb2d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.555013] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 11adaee5-8e56-4679-8373-5d5690a44ca9/11adaee5-8e56-4679-8373-5d5690a44ca9.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 740.555013] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb587abe-1c45-47dc-b5d9-a30c0e5ef425 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.578388] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for the task: (returnval){ [ 740.578388] env[61857]: value = "task-950748" [ 740.578388] env[61857]: _type = "Task" [ 740.578388] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.588476] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950748, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.627284] env[61857]: DEBUG nova.network.neutron [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Successfully updated port: 289c5ef4-b619-4720-bfc8-58a7b733beee {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 740.641340] env[61857]: DEBUG oslo_vmware.api [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950747, 'name': PowerOffVM_Task, 'duration_secs': 0.128853} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.641340] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 740.641340] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 740.641340] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b14f3d63-eedb-4df2-9270-2790cf32a9c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.666666] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 740.666859] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 740.666911] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Deleting the datastore file [datastore1] 6304714b-8988-4798-990f-d3e19774eb8f {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 740.667207] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9fed824-eefc-491b-8e5c-05d53b9478b7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.674686] env[61857]: DEBUG oslo_vmware.api [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for the task: (returnval){ [ 740.674686] env[61857]: value = "task-950750" [ 740.674686] env[61857]: _type = "Task" [ 740.674686] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.683249] env[61857]: DEBUG oslo_vmware.api [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950750, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.839166] env[61857]: DEBUG nova.compute.utils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.843038] env[61857]: DEBUG nova.compute.manager [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 740.843038] env[61857]: DEBUG nova.network.neutron [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 740.893275] env[61857]: DEBUG nova.policy [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2ef7e845007475a8a19cf3f520c3f51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1acf476aabc4166bc8505a3442367c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 740.972982] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5285af19-1f2b-3e9f-d03a-7b2b7e4b88e7, 'name': SearchDatastore_Task, 'duration_secs': 0.00941} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.974263] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.974263] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 740.974263] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.974263] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.974837] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 740.974837] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-61380248-a611-42d4-bfd1-bc2d778bb3a0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.983531] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 740.983531] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 740.983869] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89402aaf-6e38-43a9-b124-5dc373458309 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.989043] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 740.989043] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5254a18b-110a-efa1-ef68-43c9459d74a6" [ 740.989043] env[61857]: _type = "Task" [ 740.989043] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.999673] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5254a18b-110a-efa1-ef68-43c9459d74a6, 'name': SearchDatastore_Task, 'duration_secs': 0.008567} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.000462] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-251759b1-a13f-4651-a143-138403eb9331 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.005967] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 741.005967] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]527721b5-86a8-d98f-3962-e89168da04a7" [ 741.005967] env[61857]: _type = "Task" [ 741.005967] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.015231] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]527721b5-86a8-d98f-3962-e89168da04a7, 'name': SearchDatastore_Task, 'duration_secs': 0.007559} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.015471] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.015711] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] e88aed62-6561-4e29-b041-46b19cc5ef63/e88aed62-6561-4e29-b041-46b19cc5ef63.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 741.016475] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6f37020-8ec9-4dc4-9e7e-c65f29ef2ba4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.022661] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 741.022661] env[61857]: value = "task-950751" [ 741.022661] env[61857]: _type = "Task" [ 741.022661] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.032237] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.092807] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950748, 'name': ReconfigVM_Task, 'duration_secs': 0.28646} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.093196] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 11adaee5-8e56-4679-8373-5d5690a44ca9/11adaee5-8e56-4679-8373-5d5690a44ca9.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 741.093876] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2433c3ca-a6f3-48d7-b988-96f5133de050 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.103081] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for the task: (returnval){ [ 741.103081] env[61857]: value = "task-950752" [ 741.103081] env[61857]: _type = "Task" [ 741.103081] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.112222] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950752, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.132020] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "refresh_cache-c4f70321-f612-416e-b209-eb974dab9d49" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.132020] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquired lock "refresh_cache-c4f70321-f612-416e-b209-eb974dab9d49" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.132020] env[61857]: DEBUG nova.network.neutron [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.188127] env[61857]: DEBUG oslo_vmware.api [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Task: {'id': task-950750, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123184} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.188127] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 741.188127] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 741.188297] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 741.188561] env[61857]: INFO nova.compute.manager [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Took 1.07 seconds to destroy the instance on the hypervisor. [ 741.188826] env[61857]: DEBUG oslo.service.loopingcall [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.189042] env[61857]: DEBUG nova.compute.manager [-] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.189138] env[61857]: DEBUG nova.network.neutron [-] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 741.217178] env[61857]: DEBUG nova.network.neutron [-] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.292762] env[61857]: DEBUG nova.network.neutron [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Successfully created port: 527bd220-3cfd-4df1-b3c6-014e4bd05cc9 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.343641] env[61857]: DEBUG nova.compute.manager [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 741.466232] env[61857]: DEBUG nova.network.neutron [req-528082fe-dcf3-4023-8852-23efe34438fc req-f11dcbd6-6f98-44d4-a316-850d1274aaf5 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Updated VIF entry in instance network info cache for port 03542c83-4268-49c5-a829-9d7750f2a70a. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 741.466600] env[61857]: DEBUG nova.network.neutron [req-528082fe-dcf3-4023-8852-23efe34438fc req-f11dcbd6-6f98-44d4-a316-850d1274aaf5 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Updating instance_info_cache with network_info: [{"id": "03542c83-4268-49c5-a829-9d7750f2a70a", "address": "fa:16:3e:03:ac:df", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03542c83-42", "ovs_interfaceid": "03542c83-4268-49c5-a829-9d7750f2a70a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.538743] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950751, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.614207] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950752, 'name': Rename_Task, 'duration_secs': 0.295705} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.614557] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 741.614807] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-291fa401-221a-4458-a9e0-c8764df0ed38 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.625143] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for the task: (returnval){ [ 741.625143] env[61857]: value = "task-950753" [ 741.625143] env[61857]: _type = "Task" [ 741.625143] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.635363] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.719195] env[61857]: DEBUG nova.network.neutron [-] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.721173] env[61857]: DEBUG nova.network.neutron [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.900895] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223366b4-8637-4222-8fc9-3e56b69278c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.908853] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fd1593-37aa-4b2b-b5bc-ad746a54c5ae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.942686] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5ea4f2-6c15-49dd-bd53-2070baf559fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.952962] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e63d7e5-2d35-4208-a5e0-1dd028b9430b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.967827] env[61857]: DEBUG nova.compute.provider_tree [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.971537] env[61857]: DEBUG oslo_concurrency.lockutils [req-528082fe-dcf3-4023-8852-23efe34438fc req-f11dcbd6-6f98-44d4-a316-850d1274aaf5 service nova] Releasing lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.029764] env[61857]: DEBUG nova.network.neutron [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Updating instance_info_cache with network_info: [{"id": "289c5ef4-b619-4720-bfc8-58a7b733beee", "address": "fa:16:3e:16:60:bb", "network": {"id": "219627eb-2495-4c51-b7bb-4110f793f873", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-419755485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a845df4e867940639f4ff7b6a092e301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap289c5ef4-b6", "ovs_interfaceid": "289c5ef4-b619-4720-bfc8-58a7b733beee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.041464] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950751, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53086} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.041803] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] e88aed62-6561-4e29-b041-46b19cc5ef63/e88aed62-6561-4e29-b041-46b19cc5ef63.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 742.042036] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 742.042297] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30c4b1d1-ba6e-45df-bec0-d20e135cb10b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.050480] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 742.050480] env[61857]: value = "task-950754" [ 742.050480] env[61857]: _type = "Task" [ 742.050480] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.059697] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950754, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.137305] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.225714] env[61857]: INFO nova.compute.manager [-] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Took 1.04 seconds to deallocate network for instance. [ 742.355402] env[61857]: DEBUG nova.compute.manager [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 742.384891] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 742.385079] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 742.385245] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.385428] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 742.385578] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.385722] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 742.385927] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 742.386174] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 742.386397] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 742.386607] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 742.386845] env[61857]: DEBUG nova.virt.hardware [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 742.387742] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6548d0b-8f66-40ee-a4b7-cf7c7c47af65 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.396482] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf606cd-a28a-467b-ad6e-0be83a26d964 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.469275] env[61857]: DEBUG nova.scheduler.client.report [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.532078] env[61857]: DEBUG nova.compute.manager [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Received event network-vif-plugged-289c5ef4-b619-4720-bfc8-58a7b733beee {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 742.532344] env[61857]: DEBUG oslo_concurrency.lockutils [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] Acquiring lock "c4f70321-f612-416e-b209-eb974dab9d49-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.532609] env[61857]: DEBUG oslo_concurrency.lockutils [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] Lock "c4f70321-f612-416e-b209-eb974dab9d49-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.532784] env[61857]: DEBUG oslo_concurrency.lockutils [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] Lock "c4f70321-f612-416e-b209-eb974dab9d49-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.532951] env[61857]: DEBUG nova.compute.manager [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] No waiting events found dispatching network-vif-plugged-289c5ef4-b619-4720-bfc8-58a7b733beee {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 742.533549] env[61857]: WARNING nova.compute.manager [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Received unexpected event network-vif-plugged-289c5ef4-b619-4720-bfc8-58a7b733beee for instance with vm_state building and task_state spawning. [ 742.533733] env[61857]: DEBUG nova.compute.manager [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Received event network-changed-289c5ef4-b619-4720-bfc8-58a7b733beee {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 742.533938] env[61857]: DEBUG nova.compute.manager [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Refreshing instance network info cache due to event network-changed-289c5ef4-b619-4720-bfc8-58a7b733beee. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 742.534066] env[61857]: DEBUG oslo_concurrency.lockutils [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] Acquiring lock "refresh_cache-c4f70321-f612-416e-b209-eb974dab9d49" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.534966] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Releasing lock "refresh_cache-c4f70321-f612-416e-b209-eb974dab9d49" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.535256] env[61857]: DEBUG nova.compute.manager [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Instance network_info: |[{"id": "289c5ef4-b619-4720-bfc8-58a7b733beee", "address": "fa:16:3e:16:60:bb", "network": {"id": "219627eb-2495-4c51-b7bb-4110f793f873", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-419755485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a845df4e867940639f4ff7b6a092e301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap289c5ef4-b6", "ovs_interfaceid": "289c5ef4-b619-4720-bfc8-58a7b733beee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 742.535507] env[61857]: DEBUG oslo_concurrency.lockutils [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] Acquired lock "refresh_cache-c4f70321-f612-416e-b209-eb974dab9d49" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.535683] env[61857]: DEBUG nova.network.neutron [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Refreshing network info cache for port 289c5ef4-b619-4720-bfc8-58a7b733beee {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 742.537211] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:60:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaba65c3-6925-4c7f-83b6-17cd1a328e27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '289c5ef4-b619-4720-bfc8-58a7b733beee', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 742.545813] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Creating folder: Project (a845df4e867940639f4ff7b6a092e301). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 742.547095] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5e6fc938-b541-44d8-8bcd-31c652600706 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.558197] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Created folder: Project (a845df4e867940639f4ff7b6a092e301) in parent group-v214027. [ 742.558392] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Creating folder: Instances. Parent ref: group-v214056. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 742.561348] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05da1a47-9fd4-41ff-a7b1-fc8660216650 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.563145] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950754, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070374} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.563388] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 742.564420] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea9aeef-ac2c-45ea-98ac-20c1be3e7392 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.587594] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] e88aed62-6561-4e29-b041-46b19cc5ef63/e88aed62-6561-4e29-b041-46b19cc5ef63.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 742.589595] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5601531b-a197-403c-8da6-04de38225715 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.605063] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Created folder: Instances in parent group-v214056. [ 742.605437] env[61857]: DEBUG oslo.service.loopingcall [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.605565] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 742.606549] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bef23e61-a031-4157-b5b3-17971013888a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.623238] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 742.623238] env[61857]: value = "task-950757" [ 742.623238] env[61857]: _type = "Task" [ 742.623238] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.630904] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 742.630904] env[61857]: value = "task-950758" [ 742.630904] env[61857]: _type = "Task" [ 742.630904] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.634477] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950757, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.640192] env[61857]: DEBUG oslo_vmware.api [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950753, 'name': PowerOnVM_Task, 'duration_secs': 0.60643} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.640845] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 742.641149] env[61857]: INFO nova.compute.manager [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Took 8.24 seconds to spawn the instance on the hypervisor. [ 742.641298] env[61857]: DEBUG nova.compute.manager [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 742.642176] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77ca530-3c0c-40c8-bd81-a10be78b20bb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.647728] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950758, 'name': CreateVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.733060] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.974690] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.641s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.975533] env[61857]: DEBUG nova.compute.manager [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 742.977959] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.945s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.979385] env[61857]: INFO nova.compute.claims [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.055389] env[61857]: DEBUG nova.network.neutron [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Successfully updated port: 527bd220-3cfd-4df1-b3c6-014e4bd05cc9 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 743.137688] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950757, 'name': ReconfigVM_Task, 'duration_secs': 0.340542} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.144955] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Reconfigured VM instance instance-0000002c to attach disk [datastore2] e88aed62-6561-4e29-b041-46b19cc5ef63/e88aed62-6561-4e29-b041-46b19cc5ef63.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 743.147991] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65bf44cf-53fa-4d37-afcc-9faa21e62824 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.156744] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950758, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.158966] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 743.158966] env[61857]: value = "task-950759" [ 743.158966] env[61857]: _type = "Task" [ 743.158966] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.176292] env[61857]: INFO nova.compute.manager [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Took 26.95 seconds to build instance. [ 743.181748] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950759, 'name': Rename_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.437437] env[61857]: DEBUG nova.network.neutron [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Updated VIF entry in instance network info cache for port 289c5ef4-b619-4720-bfc8-58a7b733beee. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 743.437822] env[61857]: DEBUG nova.network.neutron [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Updating instance_info_cache with network_info: [{"id": "289c5ef4-b619-4720-bfc8-58a7b733beee", "address": "fa:16:3e:16:60:bb", "network": {"id": "219627eb-2495-4c51-b7bb-4110f793f873", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-419755485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a845df4e867940639f4ff7b6a092e301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap289c5ef4-b6", "ovs_interfaceid": "289c5ef4-b619-4720-bfc8-58a7b733beee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.491011] env[61857]: DEBUG nova.compute.utils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 743.492420] env[61857]: DEBUG nova.compute.manager [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Not allocating networking since 'none' was specified. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 743.558880] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "refresh_cache-18d9307d-6a20-47b4-b4e9-176f9a8c33cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.559055] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "refresh_cache-18d9307d-6a20-47b4-b4e9-176f9a8c33cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.559214] env[61857]: DEBUG nova.network.neutron [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 743.647845] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950758, 'name': CreateVM_Task, 'duration_secs': 0.558401} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.651018] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 743.651018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.651018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.651018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 743.651018] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdb5ae77-5ff0-4cce-90f0-42394b0d4d17 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.654889] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for the task: (returnval){ [ 743.654889] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520bfa9e-d696-457c-dfd1-45ff6576ffd1" [ 743.654889] env[61857]: _type = "Task" [ 743.654889] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.665440] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]520bfa9e-d696-457c-dfd1-45ff6576ffd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.671244] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950759, 'name': Rename_Task, 'duration_secs': 0.143405} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.671244] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 743.671244] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e02d615d-a3d9-4997-a6fe-a3d657686ad5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.679339] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 743.679339] env[61857]: value = "task-950760" [ 743.679339] env[61857]: _type = "Task" [ 743.679339] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.683761] env[61857]: DEBUG oslo_concurrency.lockutils [None req-516af306-c680-4833-9eab-039c99229a20 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lock "11adaee5-8e56-4679-8373-5d5690a44ca9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.789s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.690954] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950760, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.940739] env[61857]: DEBUG oslo_concurrency.lockutils [req-2698b898-e04d-4626-b8e5-c03e9c152329 req-286b7c1a-39bb-4be2-9ce8-717ff9d5365d service nova] Releasing lock "refresh_cache-c4f70321-f612-416e-b209-eb974dab9d49" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.996599] env[61857]: DEBUG nova.compute.manager [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 744.104245] env[61857]: DEBUG nova.network.neutron [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.117533] env[61857]: DEBUG nova.compute.manager [req-13d4c310-232f-4904-bec6-2e0e0fd80e9f req-c6264ec0-2868-4365-957a-eff1fd55c544 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Received event network-changed-e656ed86-dc81-4383-8dcd-254c4a462021 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 744.117729] env[61857]: DEBUG nova.compute.manager [req-13d4c310-232f-4904-bec6-2e0e0fd80e9f req-c6264ec0-2868-4365-957a-eff1fd55c544 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Refreshing instance network info cache due to event network-changed-e656ed86-dc81-4383-8dcd-254c4a462021. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 744.117989] env[61857]: DEBUG oslo_concurrency.lockutils [req-13d4c310-232f-4904-bec6-2e0e0fd80e9f req-c6264ec0-2868-4365-957a-eff1fd55c544 service nova] Acquiring lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.119064] env[61857]: DEBUG oslo_concurrency.lockutils [req-13d4c310-232f-4904-bec6-2e0e0fd80e9f req-c6264ec0-2868-4365-957a-eff1fd55c544 service nova] Acquired lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.119064] env[61857]: DEBUG nova.network.neutron [req-13d4c310-232f-4904-bec6-2e0e0fd80e9f req-c6264ec0-2868-4365-957a-eff1fd55c544 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Refreshing network info cache for port e656ed86-dc81-4383-8dcd-254c4a462021 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 744.167182] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]520bfa9e-d696-457c-dfd1-45ff6576ffd1, 'name': SearchDatastore_Task, 'duration_secs': 0.019484} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.170687] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.170687] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 744.170687] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.170687] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.170919] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 744.171048] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87173a0c-0029-42b9-ade8-6eb26a9fa525 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.180658] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 744.180834] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 744.186974] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c32fa1c-41a8-4672-967d-f869308052a8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.192844] env[61857]: DEBUG nova.compute.manager [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.201159] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950760, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.202520] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for the task: (returnval){ [ 744.202520] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52d3ba6b-0b05-32a9-c574-41d3056c1735" [ 744.202520] env[61857]: _type = "Task" [ 744.202520] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.216857] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52d3ba6b-0b05-32a9-c574-41d3056c1735, 'name': SearchDatastore_Task, 'duration_secs': 0.009739} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.217636] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bda13cc-4b89-4fa1-b68e-ecfc325d5702 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.222969] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for the task: (returnval){ [ 744.222969] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5292c13a-b83d-d88d-bc92-c4f4af1b8c44" [ 744.222969] env[61857]: _type = "Task" [ 744.222969] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.232459] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5292c13a-b83d-d88d-bc92-c4f4af1b8c44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.290032] env[61857]: DEBUG nova.network.neutron [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Updating instance_info_cache with network_info: [{"id": "527bd220-3cfd-4df1-b3c6-014e4bd05cc9", "address": "fa:16:3e:16:4d:0c", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap527bd220-3c", "ovs_interfaceid": "527bd220-3cfd-4df1-b3c6-014e4bd05cc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.486895] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a9d051-29a5-4149-ba4a-4bbacc12d4c9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.494737] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e88a1e54-afb3-4e40-ac2f-f93c979d47d4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.533235] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b89cbc9-8820-4b30-ae1a-5e0038fe6205 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.538008] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9f9628-d8a0-43df-a929-91da6daaf9d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.552234] env[61857]: DEBUG nova.compute.provider_tree [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.613692] env[61857]: DEBUG nova.compute.manager [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Received event network-vif-plugged-527bd220-3cfd-4df1-b3c6-014e4bd05cc9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 744.613935] env[61857]: DEBUG oslo_concurrency.lockutils [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] Acquiring lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.614479] env[61857]: DEBUG oslo_concurrency.lockutils [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] Lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.614479] env[61857]: DEBUG oslo_concurrency.lockutils [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] Lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.614479] env[61857]: DEBUG nova.compute.manager [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] No waiting events found dispatching network-vif-plugged-527bd220-3cfd-4df1-b3c6-014e4bd05cc9 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 744.614633] env[61857]: WARNING nova.compute.manager [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Received unexpected event network-vif-plugged-527bd220-3cfd-4df1-b3c6-014e4bd05cc9 for instance with vm_state building and task_state spawning. [ 744.614761] env[61857]: DEBUG nova.compute.manager [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Received event network-changed-527bd220-3cfd-4df1-b3c6-014e4bd05cc9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 744.614912] env[61857]: DEBUG nova.compute.manager [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Refreshing instance network info cache due to event network-changed-527bd220-3cfd-4df1-b3c6-014e4bd05cc9. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 744.616898] env[61857]: DEBUG oslo_concurrency.lockutils [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] Acquiring lock "refresh_cache-18d9307d-6a20-47b4-b4e9-176f9a8c33cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.698423] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950760, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.717045] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.735756] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5292c13a-b83d-d88d-bc92-c4f4af1b8c44, 'name': SearchDatastore_Task, 'duration_secs': 0.008759} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.739022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.739022] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] c4f70321-f612-416e-b209-eb974dab9d49/c4f70321-f612-416e-b209-eb974dab9d49.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 744.739022] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5deaa1ed-0af1-43d7-af65-1064acb806ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.745388] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for the task: (returnval){ [ 744.745388] env[61857]: value = "task-950761" [ 744.745388] env[61857]: _type = "Task" [ 744.745388] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.753673] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950761, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.793619] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "refresh_cache-18d9307d-6a20-47b4-b4e9-176f9a8c33cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.793962] env[61857]: DEBUG nova.compute.manager [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Instance network_info: |[{"id": "527bd220-3cfd-4df1-b3c6-014e4bd05cc9", "address": "fa:16:3e:16:4d:0c", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap527bd220-3c", "ovs_interfaceid": "527bd220-3cfd-4df1-b3c6-014e4bd05cc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 744.794293] env[61857]: DEBUG oslo_concurrency.lockutils [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] Acquired lock "refresh_cache-18d9307d-6a20-47b4-b4e9-176f9a8c33cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.794485] env[61857]: DEBUG nova.network.neutron [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Refreshing network info cache for port 527bd220-3cfd-4df1-b3c6-014e4bd05cc9 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 744.795871] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:4d:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '527bd220-3cfd-4df1-b3c6-014e4bd05cc9', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 744.804238] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Creating folder: Project (b1acf476aabc4166bc8505a3442367c6). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 744.804870] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aaaff11c-358b-49e0-85da-bb8752b3a09b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.821201] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Created folder: Project (b1acf476aabc4166bc8505a3442367c6) in parent group-v214027. [ 744.821417] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Creating folder: Instances. Parent ref: group-v214059. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 744.821727] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a416c673-45c0-4442-94a3-fc3ddeba061f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.831085] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Created folder: Instances in parent group-v214059. [ 744.832148] env[61857]: DEBUG oslo.service.loopingcall [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.832148] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 744.832148] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ff234eb-55d5-4653-a9f3-25ab034e6748 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.860528] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 744.860528] env[61857]: value = "task-950764" [ 744.860528] env[61857]: _type = "Task" [ 744.860528] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.870855] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950764, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.036482] env[61857]: DEBUG nova.compute.manager [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.055137] env[61857]: DEBUG nova.scheduler.client.report [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.064457] env[61857]: DEBUG nova.network.neutron [req-13d4c310-232f-4904-bec6-2e0e0fd80e9f req-c6264ec0-2868-4365-957a-eff1fd55c544 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Updated VIF entry in instance network info cache for port e656ed86-dc81-4383-8dcd-254c4a462021. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 745.065493] env[61857]: DEBUG nova.network.neutron [req-13d4c310-232f-4904-bec6-2e0e0fd80e9f req-c6264ec0-2868-4365-957a-eff1fd55c544 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Updating instance_info_cache with network_info: [{"id": "e656ed86-dc81-4383-8dcd-254c4a462021", "address": "fa:16:3e:14:73:a6", "network": {"id": "32c36dba-b78e-4e41-ae8c-40f4b673555f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1785406565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e0848df81fc84d25a4bb79d6f0b4b7cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape656ed86-dc", "ovs_interfaceid": "e656ed86-dc81-4383-8dcd-254c4a462021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.078751] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.079029] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.079293] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.079359] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.079502] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.079652] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.079865] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.080028] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.080300] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.080395] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.080772] env[61857]: DEBUG nova.virt.hardware [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.082324] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5afdf11-74a3-4793-9943-e5fc2f03af9a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.094498] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d08eb25-e7c7-4cf6-af30-f49e0be383f9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.111122] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 745.117582] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Creating folder: Project (c724979e84ea48b8ba69e5cdab6f5f23). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 745.117956] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e50e7cd-db69-44d9-aed1-438787e4a41d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.128514] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Created folder: Project (c724979e84ea48b8ba69e5cdab6f5f23) in parent group-v214027. [ 745.128718] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Creating folder: Instances. Parent ref: group-v214062. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 745.130130] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-256faf3a-5f2d-4fa5-b7a2-dc5ff95477a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.139875] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Created folder: Instances in parent group-v214062. [ 745.142953] env[61857]: DEBUG oslo.service.loopingcall [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.142953] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 745.142953] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7097bff7-0526-48e1-85bd-97d57db96a56 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.159072] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 745.159072] env[61857]: value = "task-950767" [ 745.159072] env[61857]: _type = "Task" [ 745.159072] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.170802] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950767, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.196032] env[61857]: DEBUG oslo_vmware.api [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-950760, 'name': PowerOnVM_Task, 'duration_secs': 1.204738} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.196032] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 745.196032] env[61857]: INFO nova.compute.manager [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Took 8.20 seconds to spawn the instance on the hypervisor. [ 745.196032] env[61857]: DEBUG nova.compute.manager [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 745.196032] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9535f3ef-b6e1-4b99-8fd1-5bb7c5474b26 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.256051] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950761, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.371868] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950764, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.563578] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.564139] env[61857]: DEBUG nova.compute.manager [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 745.570369] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.014s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.572048] env[61857]: INFO nova.compute.claims [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.575620] env[61857]: DEBUG oslo_concurrency.lockutils [req-13d4c310-232f-4904-bec6-2e0e0fd80e9f req-c6264ec0-2868-4365-957a-eff1fd55c544 service nova] Releasing lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.669950] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950767, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.724874] env[61857]: INFO nova.compute.manager [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Took 28.06 seconds to build instance. [ 745.760099] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950761, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547629} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.760099] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] c4f70321-f612-416e-b209-eb974dab9d49/c4f70321-f612-416e-b209-eb974dab9d49.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 745.760349] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 745.760691] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca28471c-4544-435e-bde2-7aecf42c7d8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.773405] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for the task: (returnval){ [ 745.773405] env[61857]: value = "task-950768" [ 745.773405] env[61857]: _type = "Task" [ 745.773405] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.780819] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.828531] env[61857]: DEBUG nova.network.neutron [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Updated VIF entry in instance network info cache for port 527bd220-3cfd-4df1-b3c6-014e4bd05cc9. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 745.828531] env[61857]: DEBUG nova.network.neutron [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Updating instance_info_cache with network_info: [{"id": "527bd220-3cfd-4df1-b3c6-014e4bd05cc9", "address": "fa:16:3e:16:4d:0c", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap527bd220-3c", "ovs_interfaceid": "527bd220-3cfd-4df1-b3c6-014e4bd05cc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.871382] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950764, 'name': CreateVM_Task, 'duration_secs': 0.565761} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.872127] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 745.872959] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.873464] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.873923] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 745.874312] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82f22176-5fc3-4d8a-9401-f4a5f72393fb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.879562] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 745.879562] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]527244f5-5d01-8a19-b11e-217bf9018e08" [ 745.879562] env[61857]: _type = "Task" [ 745.879562] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.890478] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]527244f5-5d01-8a19-b11e-217bf9018e08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.077604] env[61857]: DEBUG nova.compute.utils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 746.082141] env[61857]: DEBUG nova.compute.manager [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Not allocating networking since 'none' was specified. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 746.175371] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950767, 'name': CreateVM_Task, 'duration_secs': 0.692547} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.175616] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 746.175958] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.227211] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c99bd1e-f105-4146-80e4-cfd8b86dcadb tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "e88aed62-6561-4e29-b041-46b19cc5ef63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.334s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.282094] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.219853} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.282459] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 746.283325] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4bdc325-3ef4-4753-91da-973a4d8c4f95 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.305886] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] c4f70321-f612-416e-b209-eb974dab9d49/c4f70321-f612-416e-b209-eb974dab9d49.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 746.306510] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ef68bb3-6b3d-448f-90c3-91a45ad9e3b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.326963] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for the task: (returnval){ [ 746.326963] env[61857]: value = "task-950769" [ 746.326963] env[61857]: _type = "Task" [ 746.326963] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.330467] env[61857]: DEBUG oslo_concurrency.lockutils [req-dab6e7ce-9559-4146-b953-a2d19965ffe8 req-d09bddfc-3a55-49b9-82f8-b8240115dc5f service nova] Releasing lock "refresh_cache-18d9307d-6a20-47b4-b4e9-176f9a8c33cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.335651] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950769, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.392058] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]527244f5-5d01-8a19-b11e-217bf9018e08, 'name': SearchDatastore_Task, 'duration_secs': 0.016969} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.392318] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.392599] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 746.392884] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.393045] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.393522] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 746.393812] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.394133] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 746.394457] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4bfe4a6a-8861-4c8a-9680-9237ece647ae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.396352] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0d1af14-c603-4d24-abe2-2daff0ea5306 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.401661] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 746.401661] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b6304a-6b9f-b6e6-9be6-cdaddd4c03c6" [ 746.401661] env[61857]: _type = "Task" [ 746.401661] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.408718] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 746.408906] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 746.412506] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9156503e-032e-4e72-afd8-d6b14293680f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.414874] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b6304a-6b9f-b6e6-9be6-cdaddd4c03c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.417726] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 746.417726] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b46a53-d1e0-9eb6-67eb-b95ffb8e1c1e" [ 746.417726] env[61857]: _type = "Task" [ 746.417726] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.426520] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b46a53-d1e0-9eb6-67eb-b95ffb8e1c1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.586412] env[61857]: DEBUG nova.compute.manager [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 746.679666] env[61857]: DEBUG nova.compute.manager [req-96d7151a-db82-49ab-abed-ff3cbc2379f0 req-adb7e5db-5f4b-4fd1-9bb9-3b804ed0e236 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Received event network-changed-03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 746.679865] env[61857]: DEBUG nova.compute.manager [req-96d7151a-db82-49ab-abed-ff3cbc2379f0 req-adb7e5db-5f4b-4fd1-9bb9-3b804ed0e236 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Refreshing instance network info cache due to event network-changed-03542c83-4268-49c5-a829-9d7750f2a70a. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 746.680099] env[61857]: DEBUG oslo_concurrency.lockutils [req-96d7151a-db82-49ab-abed-ff3cbc2379f0 req-adb7e5db-5f4b-4fd1-9bb9-3b804ed0e236 service nova] Acquiring lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.680284] env[61857]: DEBUG oslo_concurrency.lockutils [req-96d7151a-db82-49ab-abed-ff3cbc2379f0 req-adb7e5db-5f4b-4fd1-9bb9-3b804ed0e236 service nova] Acquired lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.680465] env[61857]: DEBUG nova.network.neutron [req-96d7151a-db82-49ab-abed-ff3cbc2379f0 req-adb7e5db-5f4b-4fd1-9bb9-3b804ed0e236 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Refreshing network info cache for port 03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 746.732167] env[61857]: DEBUG nova.compute.manager [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.837096] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950769, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.914745] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b6304a-6b9f-b6e6-9be6-cdaddd4c03c6, 'name': SearchDatastore_Task, 'duration_secs': 0.017974} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.914745] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.914959] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 746.915301] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.927813] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b46a53-d1e0-9eb6-67eb-b95ffb8e1c1e, 'name': SearchDatastore_Task, 'duration_secs': 0.013091} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.930790] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5ada972-a807-4687-ada1-002234fae812 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.935516] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 746.935516] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad4c15-21f3-3cf9-6c5d-2d45dddec50d" [ 746.935516] env[61857]: _type = "Task" [ 746.935516] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.944060] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad4c15-21f3-3cf9-6c5d-2d45dddec50d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.021156] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf81760-8874-4927-aea3-4a5baad31887 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.029504] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1856e0a4-fa41-486e-8d3c-79c27d4853f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.063047] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3afac7e-f034-4b0b-a3cd-ebdbd335d4d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.070766] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90db8802-c609-4f0b-92f5-784299d01b25 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.084569] env[61857]: DEBUG nova.compute.provider_tree [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.253597] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.337890] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950769, 'name': ReconfigVM_Task, 'duration_secs': 0.831384} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.340522] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Reconfigured VM instance instance-0000002d to attach disk [datastore2] c4f70321-f612-416e-b209-eb974dab9d49/c4f70321-f612-416e-b209-eb974dab9d49.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 747.342103] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14210b60-0f80-4edf-81be-c7106f85adb9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.347735] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for the task: (returnval){ [ 747.347735] env[61857]: value = "task-950770" [ 747.347735] env[61857]: _type = "Task" [ 747.347735] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.358630] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950770, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.447106] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad4c15-21f3-3cf9-6c5d-2d45dddec50d, 'name': SearchDatastore_Task, 'duration_secs': 0.012841} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.447163] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.447392] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 18d9307d-6a20-47b4-b4e9-176f9a8c33cc/18d9307d-6a20-47b4-b4e9-176f9a8c33cc.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 747.447669] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.447900] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 747.448148] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5eb4efe-9b40-412c-8b39-a49ceddb06c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.450678] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-319d9149-e5c1-4ed0-856f-8ec146c90347 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.458975] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 747.458975] env[61857]: value = "task-950771" [ 747.458975] env[61857]: _type = "Task" [ 747.458975] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.460794] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 747.460962] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 747.462223] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3bcebb7-f394-415a-bab3-328aa3b46d7b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.468985] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950771, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.472786] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 747.472786] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524782c6-9ac1-5e81-6b8d-db7a19233771" [ 747.472786] env[61857]: _type = "Task" [ 747.472786] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.473662] env[61857]: DEBUG nova.network.neutron [req-96d7151a-db82-49ab-abed-ff3cbc2379f0 req-adb7e5db-5f4b-4fd1-9bb9-3b804ed0e236 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Updated VIF entry in instance network info cache for port 03542c83-4268-49c5-a829-9d7750f2a70a. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 747.473977] env[61857]: DEBUG nova.network.neutron [req-96d7151a-db82-49ab-abed-ff3cbc2379f0 req-adb7e5db-5f4b-4fd1-9bb9-3b804ed0e236 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Updating instance_info_cache with network_info: [{"id": "03542c83-4268-49c5-a829-9d7750f2a70a", "address": "fa:16:3e:03:ac:df", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03542c83-42", "ovs_interfaceid": "03542c83-4268-49c5-a829-9d7750f2a70a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.488307] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524782c6-9ac1-5e81-6b8d-db7a19233771, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.588894] env[61857]: DEBUG nova.scheduler.client.report [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.600591] env[61857]: DEBUG nova.compute.manager [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 747.635213] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 747.635213] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 747.635213] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.635213] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 747.635717] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.635717] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 747.635717] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 747.635717] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 747.635717] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 747.635973] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 747.635973] env[61857]: DEBUG nova.virt.hardware [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 747.636739] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe1a247-30db-49d3-88e0-d3053376a897 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.645675] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394af1da-9939-43cc-97bd-1c0fae168e36 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.660364] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 747.667138] env[61857]: DEBUG oslo.service.loopingcall [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.667507] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 747.667737] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3602d059-96f4-42f4-a280-e22a831982bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.687456] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 747.687456] env[61857]: value = "task-950772" [ 747.687456] env[61857]: _type = "Task" [ 747.687456] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.699852] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950772, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.859211] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950770, 'name': Rename_Task, 'duration_secs': 0.165507} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.859522] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 747.859771] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8aa499bd-baeb-4f66-b075-2f66071378a3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.867294] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for the task: (returnval){ [ 747.867294] env[61857]: value = "task-950773" [ 747.867294] env[61857]: _type = "Task" [ 747.867294] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.876882] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950773, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.966646] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950771, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491242} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.967339] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 18d9307d-6a20-47b4-b4e9-176f9a8c33cc/18d9307d-6a20-47b4-b4e9-176f9a8c33cc.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 747.967339] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 747.967530] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b9765262-e992-4813-996e-217e24aae677 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.975396] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 747.975396] env[61857]: value = "task-950774" [ 747.975396] env[61857]: _type = "Task" [ 747.975396] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.982308] env[61857]: DEBUG oslo_concurrency.lockutils [req-96d7151a-db82-49ab-abed-ff3cbc2379f0 req-adb7e5db-5f4b-4fd1-9bb9-3b804ed0e236 service nova] Releasing lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.992225] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950774, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.992225] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524782c6-9ac1-5e81-6b8d-db7a19233771, 'name': SearchDatastore_Task, 'duration_secs': 0.020042} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.992582] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05ae9f06-5243-4fa2-9cb5-e07f42d67b13 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.997987] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 747.997987] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525570ad-7959-4520-8a3a-2748cc5084aa" [ 747.997987] env[61857]: _type = "Task" [ 747.997987] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.005896] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525570ad-7959-4520-8a3a-2748cc5084aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.093680] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.094308] env[61857]: DEBUG nova.compute.manager [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.097123] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.508s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.098545] env[61857]: INFO nova.compute.claims [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.201314] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950772, 'name': CreateVM_Task, 'duration_secs': 0.436569} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.202055] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 748.202233] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.202573] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.202957] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 748.203512] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-754ff2b1-570d-4a68-9ad0-36fb440239ca {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.209890] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 748.209890] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]528a1eac-8ea8-5de5-5dec-e0c2c880c71f" [ 748.209890] env[61857]: _type = "Task" [ 748.209890] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.221504] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528a1eac-8ea8-5de5-5dec-e0c2c880c71f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.378376] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950773, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.485107] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950774, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063794} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.485776] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 748.486606] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287aeb48-999a-42d2-b691-c0a61dcb4692 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.512159] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] 18d9307d-6a20-47b4-b4e9-176f9a8c33cc/18d9307d-6a20-47b4-b4e9-176f9a8c33cc.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 748.515405] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69e8db4e-133b-47db-ba03-9d8eb37f4bda {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.537471] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525570ad-7959-4520-8a3a-2748cc5084aa, 'name': SearchDatastore_Task, 'duration_secs': 0.010513} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.538682] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.538938] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] d49008e6-7789-4785-8eea-9b3fa36c4ccb/d49008e6-7789-4785-8eea-9b3fa36c4ccb.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 748.539266] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 748.539266] env[61857]: value = "task-950775" [ 748.539266] env[61857]: _type = "Task" [ 748.539266] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.539452] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4dedf639-7704-4857-bef7-4c6d989c9cea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.549499] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950775, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.550770] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 748.550770] env[61857]: value = "task-950776" [ 748.550770] env[61857]: _type = "Task" [ 748.550770] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.558387] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950776, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.603161] env[61857]: DEBUG nova.compute.utils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.606604] env[61857]: DEBUG nova.compute.manager [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Not allocating networking since 'none' was specified. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 748.722786] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528a1eac-8ea8-5de5-5dec-e0c2c880c71f, 'name': SearchDatastore_Task, 'duration_secs': 0.079571} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.723268] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.723671] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 748.724015] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.724292] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.724598] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 748.724976] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99853896-f98a-470a-8310-6055a076684e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.735839] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 748.736138] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 748.737234] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-713da2c6-7ca2-44e1-b5ed-a5d4266fe545 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.743067] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 748.743067] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52cdcb03-2d99-aad9-73bd-a5bb63c1f665" [ 748.743067] env[61857]: _type = "Task" [ 748.743067] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.751393] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52cdcb03-2d99-aad9-73bd-a5bb63c1f665, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.769932] env[61857]: DEBUG nova.compute.manager [req-1e321c50-9d12-4018-b47d-706974ad3cdc req-b1960ae9-f82e-4b70-b4ae-3b716b9dbd51 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Received event network-changed-03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 748.770232] env[61857]: DEBUG nova.compute.manager [req-1e321c50-9d12-4018-b47d-706974ad3cdc req-b1960ae9-f82e-4b70-b4ae-3b716b9dbd51 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Refreshing instance network info cache due to event network-changed-03542c83-4268-49c5-a829-9d7750f2a70a. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 748.770530] env[61857]: DEBUG oslo_concurrency.lockutils [req-1e321c50-9d12-4018-b47d-706974ad3cdc req-b1960ae9-f82e-4b70-b4ae-3b716b9dbd51 service nova] Acquiring lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.770749] env[61857]: DEBUG oslo_concurrency.lockutils [req-1e321c50-9d12-4018-b47d-706974ad3cdc req-b1960ae9-f82e-4b70-b4ae-3b716b9dbd51 service nova] Acquired lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.770981] env[61857]: DEBUG nova.network.neutron [req-1e321c50-9d12-4018-b47d-706974ad3cdc req-b1960ae9-f82e-4b70-b4ae-3b716b9dbd51 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Refreshing network info cache for port 03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 748.879329] env[61857]: DEBUG oslo_vmware.api [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950773, 'name': PowerOnVM_Task, 'duration_secs': 0.946556} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.879604] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 748.879818] env[61857]: INFO nova.compute.manager [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Took 9.18 seconds to spawn the instance on the hypervisor. [ 748.880089] env[61857]: DEBUG nova.compute.manager [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 748.880948] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b58018-9435-4d64-a7d3-70083c219cc3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.051856] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950775, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.060876] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950776, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.107811] env[61857]: DEBUG nova.compute.manager [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.255839] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52cdcb03-2d99-aad9-73bd-a5bb63c1f665, 'name': SearchDatastore_Task, 'duration_secs': 0.0154} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.260011] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-005fa717-e728-4647-bc22-37c42d665ddc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.269068] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 749.269068] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e057aa-0cce-7030-0299-ced49cf5363f" [ 749.269068] env[61857]: _type = "Task" [ 749.269068] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.279433] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e057aa-0cce-7030-0299-ced49cf5363f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.400036] env[61857]: INFO nova.compute.manager [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Took 29.82 seconds to build instance. [ 749.564935] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950775, 'name': ReconfigVM_Task, 'duration_secs': 0.931531} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.568793] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Reconfigured VM instance instance-0000002e to attach disk [datastore1] 18d9307d-6a20-47b4-b4e9-176f9a8c33cc/18d9307d-6a20-47b4-b4e9-176f9a8c33cc.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 749.572554] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e45d1d22-ba0a-4bd3-8442-88897c8d4b21 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.575868] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950776, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.808927} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.576637] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] d49008e6-7789-4785-8eea-9b3fa36c4ccb/d49008e6-7789-4785-8eea-9b3fa36c4ccb.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 749.576995] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 749.577357] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-26d3c4f2-a791-4322-a69d-71d801d69041 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.581703] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba57700-d5d3-4694-a430-e561b3d8f110 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.587495] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 749.587495] env[61857]: value = "task-950777" [ 749.587495] env[61857]: _type = "Task" [ 749.587495] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.594119] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 749.594119] env[61857]: value = "task-950778" [ 749.594119] env[61857]: _type = "Task" [ 749.594119] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.595186] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19be8b06-62ed-494a-96f3-5e28a05deac0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.603610] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950777, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.609677] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950778, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.637130] env[61857]: DEBUG nova.network.neutron [req-1e321c50-9d12-4018-b47d-706974ad3cdc req-b1960ae9-f82e-4b70-b4ae-3b716b9dbd51 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Updated VIF entry in instance network info cache for port 03542c83-4268-49c5-a829-9d7750f2a70a. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 749.637473] env[61857]: DEBUG nova.network.neutron [req-1e321c50-9d12-4018-b47d-706974ad3cdc req-b1960ae9-f82e-4b70-b4ae-3b716b9dbd51 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Updating instance_info_cache with network_info: [{"id": "03542c83-4268-49c5-a829-9d7750f2a70a", "address": "fa:16:3e:03:ac:df", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03542c83-42", "ovs_interfaceid": "03542c83-4268-49c5-a829-9d7750f2a70a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.638931] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc3f2e17-acf7-4ef9-8df6-cd06d6096fd4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.648336] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36c6c40-9a4d-4628-b4e4-d14aeffac26c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.664592] env[61857]: DEBUG nova.compute.provider_tree [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.786360] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e057aa-0cce-7030-0299-ced49cf5363f, 'name': SearchDatastore_Task, 'duration_secs': 0.051499} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.786360] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.786360] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 3cacef7b-dd24-4d33-9500-bbff03e342bb/3cacef7b-dd24-4d33-9500-bbff03e342bb.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 749.787562] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-11a2f6b2-2a03-43f6-bf77-939a974939be {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.796910] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 749.796910] env[61857]: value = "task-950779" [ 749.796910] env[61857]: _type = "Task" [ 749.796910] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.811032] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950779, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.902441] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4efdec93-8788-4c57-8481-21b35f5b64d2 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "c4f70321-f612-416e-b209-eb974dab9d49" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.799s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.098579] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950777, 'name': Rename_Task, 'duration_secs': 0.297131} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.101690] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 750.101943] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-232571bf-1499-4da4-b02d-b7b01aabe173 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.108022] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950778, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.178641} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.109531] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 750.109531] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 750.109531] env[61857]: value = "task-950780" [ 750.109531] env[61857]: _type = "Task" [ 750.109531] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.110177] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f924e46b-0e65-47eb-a493-37845a8200fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.132260] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] d49008e6-7789-4785-8eea-9b3fa36c4ccb/d49008e6-7789-4785-8eea-9b3fa36c4ccb.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 750.135373] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89f58c99-0cd1-4555-ad92-f884cdde668e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.149258] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.150370] env[61857]: DEBUG nova.compute.manager [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.152831] env[61857]: DEBUG oslo_concurrency.lockutils [req-1e321c50-9d12-4018-b47d-706974ad3cdc req-b1960ae9-f82e-4b70-b4ae-3b716b9dbd51 service nova] Releasing lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.162362] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 750.162362] env[61857]: value = "task-950781" [ 750.162362] env[61857]: _type = "Task" [ 750.162362] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.167913] env[61857]: DEBUG nova.scheduler.client.report [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.180227] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950781, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.188953] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.189220] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.189405] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.189603] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.189753] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.189910] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.190208] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.190377] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.190553] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.190720] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.190924] env[61857]: DEBUG nova.virt.hardware [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.192560] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c6e0b0-6e0c-454f-8c83-c2aaf68d2d35 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.203306] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1c3da9-563e-4736-9508-003f136958d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.220102] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 750.226366] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Creating folder: Project (aca60994379748829a013d8c49aa4f0c). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 750.227146] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4814384e-6bef-4dff-ad83-164e5fc56a7b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.239093] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Created folder: Project (aca60994379748829a013d8c49aa4f0c) in parent group-v214027. [ 750.239885] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Creating folder: Instances. Parent ref: group-v214066. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 750.239885] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a304a79e-b562-47c9-9002-66057748a264 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.252172] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Created folder: Instances in parent group-v214066. [ 750.252606] env[61857]: DEBUG oslo.service.loopingcall [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 750.252844] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 750.253089] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b32449f4-4798-46c6-8167-e06d971d0bbf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.275499] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 750.275499] env[61857]: value = "task-950784" [ 750.275499] env[61857]: _type = "Task" [ 750.275499] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.286288] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950784, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.308650] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950779, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.356221] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "c27f6f5b-6423-4eee-be53-dc48c93df893" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.356522] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "c27f6f5b-6423-4eee-be53-dc48c93df893" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.405767] env[61857]: DEBUG nova.compute.manager [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.622727] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950780, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.671480] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950781, 'name': ReconfigVM_Task, 'duration_secs': 0.493449} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.671758] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Reconfigured VM instance instance-0000002f to attach disk [datastore1] d49008e6-7789-4785-8eea-9b3fa36c4ccb/d49008e6-7789-4785-8eea-9b3fa36c4ccb.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 750.672388] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7195564-70ed-42a1-ac09-0c871b6ec63b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.679305] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 750.679305] env[61857]: value = "task-950785" [ 750.679305] env[61857]: _type = "Task" [ 750.679305] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.683035] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.683140] env[61857]: DEBUG nova.compute.manager [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.685946] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.249s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.694718] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950785, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.785112] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950784, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.807828] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950779, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677535} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.808970] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 3cacef7b-dd24-4d33-9500-bbff03e342bb/3cacef7b-dd24-4d33-9500-bbff03e342bb.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 750.808970] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 750.808970] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21da2138-e2f9-4be4-a4ca-6c575dae923e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.815590] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 750.815590] env[61857]: value = "task-950786" [ 750.815590] env[61857]: _type = "Task" [ 750.815590] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.824896] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950786, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.931343] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.123099] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950780, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.191118] env[61857]: DEBUG nova.compute.utils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.199087] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950785, 'name': Rename_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.199087] env[61857]: DEBUG nova.compute.manager [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 751.199087] env[61857]: DEBUG nova.network.neutron [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 751.252256] env[61857]: DEBUG nova.policy [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0fc4a0ead0d5459baaca6d3d36e44d44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba507a4694a3488bab7aa929657ba9f4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 751.286028] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950784, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.327352] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950786, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111617} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.327352] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 751.328021] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a050bd84-25ca-4341-8fd8-86a3599ecaf8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.348762] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 3cacef7b-dd24-4d33-9500-bbff03e342bb/3cacef7b-dd24-4d33-9500-bbff03e342bb.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.349046] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46021071-5027-4aa8-90a7-21c1284a344e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.369629] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 751.369629] env[61857]: value = "task-950787" [ 751.369629] env[61857]: _type = "Task" [ 751.369629] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.377742] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950787, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.559967] env[61857]: DEBUG nova.network.neutron [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Successfully created port: af54a795-3d81-4edb-9d2c-05a68fc16a6f {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.623886] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950780, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.689990] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950785, 'name': Rename_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.697770] env[61857]: DEBUG nova.compute.manager [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 751.727334] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 2c187f64-8a36-4dfd-94e3-8ea944dbac24 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 751.727675] env[61857]: WARNING nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 6304714b-8988-4798-990f-d3e19774eb8f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 751.727845] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 11adaee5-8e56-4679-8373-5d5690a44ca9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 751.727983] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e88aed62-6561-4e29-b041-46b19cc5ef63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 751.728136] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c4f70321-f612-416e-b209-eb974dab9d49 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 751.729569] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 18d9307d-6a20-47b4-b4e9-176f9a8c33cc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 751.729569] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance d49008e6-7789-4785-8eea-9b3fa36c4ccb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 751.729569] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 3cacef7b-dd24-4d33-9500-bbff03e342bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 751.729569] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e286e6d0-25db-4d86-b9a2-3538bc3f624a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 751.729763] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 7a800abe-ea0a-4077-a5da-dd60eab917e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 751.787969] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950784, 'name': CreateVM_Task, 'duration_secs': 1.445515} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.788386] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 751.788854] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.789087] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.789428] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 751.789755] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd1b12a0-1bd4-4492-8f03-6f8b145c83fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.795018] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 751.795018] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523e268a-9cb6-2993-2108-942e5436139a" [ 751.795018] env[61857]: _type = "Task" [ 751.795018] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.803484] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523e268a-9cb6-2993-2108-942e5436139a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.881411] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950787, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.124770] env[61857]: DEBUG oslo_vmware.api [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-950780, 'name': PowerOnVM_Task, 'duration_secs': 1.759157} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.125057] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 752.125267] env[61857]: INFO nova.compute.manager [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Took 9.77 seconds to spawn the instance on the hypervisor. [ 752.125448] env[61857]: DEBUG nova.compute.manager [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 752.126219] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5027c2ac-5902-4df8-be6e-309783c9f5b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.190063] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950785, 'name': Rename_Task, 'duration_secs': 1.146008} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.190415] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 752.190673] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67904ad8-9d10-4b33-9dbc-bcd7bc80666d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.196783] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 752.196783] env[61857]: value = "task-950788" [ 752.196783] env[61857]: _type = "Task" [ 752.196783] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.208958] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950788, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.232079] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c4883f84-e62b-4b59-9484-270d82dc34e0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.305492] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523e268a-9cb6-2993-2108-942e5436139a, 'name': SearchDatastore_Task, 'duration_secs': 0.021709} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.305492] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.305492] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.305492] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.305798] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.305798] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.306020] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5fea64a-cb0b-466e-a99d-0a3954101c34 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.320692] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.320879] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 752.321589] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0f99578-88b1-4673-8548-2ec85da351fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.326864] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 752.326864] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b79505-6ccd-9fd2-4a28-e2e0a3dfaa4e" [ 752.326864] env[61857]: _type = "Task" [ 752.326864] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.334383] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b79505-6ccd-9fd2-4a28-e2e0a3dfaa4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.379897] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950787, 'name': ReconfigVM_Task, 'duration_secs': 0.557817} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.380263] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 3cacef7b-dd24-4d33-9500-bbff03e342bb/3cacef7b-dd24-4d33-9500-bbff03e342bb.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.381238] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e18b73e-486b-40bc-a9d2-fcf2f771c54a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.387932] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 752.387932] env[61857]: value = "task-950789" [ 752.387932] env[61857]: _type = "Task" [ 752.387932] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.396097] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950789, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.642738] env[61857]: INFO nova.compute.manager [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Took 28.59 seconds to build instance. [ 752.706588] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950788, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.711851] env[61857]: DEBUG nova.compute.manager [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.714395] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "c4f70321-f612-416e-b209-eb974dab9d49" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.714395] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "c4f70321-f612-416e-b209-eb974dab9d49" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.714566] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "c4f70321-f612-416e-b209-eb974dab9d49-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.715289] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "c4f70321-f612-416e-b209-eb974dab9d49-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.715289] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "c4f70321-f612-416e-b209-eb974dab9d49-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.716775] env[61857]: INFO nova.compute.manager [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Terminating instance [ 752.718566] env[61857]: DEBUG nova.compute.manager [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.718752] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 752.719626] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d3a0a8-7170-475a-9ccb-c289fbb713b9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.728362] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 752.728580] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5c714af-fbb2-4318-a9ea-3c0638b78416 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.735079] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 781a2790-e317-46fe-9be8-40c9e5f1f771 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.737034] env[61857]: DEBUG oslo_vmware.api [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for the task: (returnval){ [ 752.737034] env[61857]: value = "task-950790" [ 752.737034] env[61857]: _type = "Task" [ 752.737034] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.740026] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.740026] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.740026] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.740408] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.740408] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.740408] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.740408] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.740408] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.740594] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.740683] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.740925] env[61857]: DEBUG nova.virt.hardware [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.741847] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38d55af-1c96-4a70-851f-28edd534c127 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.755993] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f9c1320-b6ec-4d54-a683-6589217fea47 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.759991] env[61857]: DEBUG oslo_vmware.api [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950790, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.837352] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b79505-6ccd-9fd2-4a28-e2e0a3dfaa4e, 'name': SearchDatastore_Task, 'duration_secs': 0.017321} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.837986] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d76b3e2-6e42-47c2-9915-1aee901d642c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.843718] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 752.843718] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]529b5bbd-88e3-b2c7-4c00-e628f15286d2" [ 752.843718] env[61857]: _type = "Task" [ 752.843718] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.853543] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529b5bbd-88e3-b2c7-4c00-e628f15286d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.900158] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950789, 'name': Rename_Task, 'duration_secs': 0.193164} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.900449] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 752.900650] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff1d7e8f-9fe9-44e6-9e68-ae7166c7ee77 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.912367] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 752.912367] env[61857]: value = "task-950791" [ 752.912367] env[61857]: _type = "Task" [ 752.912367] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.918381] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950791, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.146016] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b53ab48a-d38d-4811-bc41-d521a53d88c8 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.416s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.183183] env[61857]: DEBUG nova.compute.manager [req-9d2267a6-adb1-480c-a67f-ebf7bc4cef90 req-ec84c896-9f5a-49c2-84f8-8c6c32fa039c service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Received event network-vif-plugged-af54a795-3d81-4edb-9d2c-05a68fc16a6f {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 753.183659] env[61857]: DEBUG oslo_concurrency.lockutils [req-9d2267a6-adb1-480c-a67f-ebf7bc4cef90 req-ec84c896-9f5a-49c2-84f8-8c6c32fa039c service nova] Acquiring lock "7a800abe-ea0a-4077-a5da-dd60eab917e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.183659] env[61857]: DEBUG oslo_concurrency.lockutils [req-9d2267a6-adb1-480c-a67f-ebf7bc4cef90 req-ec84c896-9f5a-49c2-84f8-8c6c32fa039c service nova] Lock "7a800abe-ea0a-4077-a5da-dd60eab917e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.184338] env[61857]: DEBUG oslo_concurrency.lockutils [req-9d2267a6-adb1-480c-a67f-ebf7bc4cef90 req-ec84c896-9f5a-49c2-84f8-8c6c32fa039c service nova] Lock "7a800abe-ea0a-4077-a5da-dd60eab917e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.184338] env[61857]: DEBUG nova.compute.manager [req-9d2267a6-adb1-480c-a67f-ebf7bc4cef90 req-ec84c896-9f5a-49c2-84f8-8c6c32fa039c service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] No waiting events found dispatching network-vif-plugged-af54a795-3d81-4edb-9d2c-05a68fc16a6f {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 753.184652] env[61857]: WARNING nova.compute.manager [req-9d2267a6-adb1-480c-a67f-ebf7bc4cef90 req-ec84c896-9f5a-49c2-84f8-8c6c32fa039c service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Received unexpected event network-vif-plugged-af54a795-3d81-4edb-9d2c-05a68fc16a6f for instance with vm_state building and task_state spawning. [ 753.207781] env[61857]: DEBUG oslo_vmware.api [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950788, 'name': PowerOnVM_Task, 'duration_secs': 0.989391} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.208088] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 753.208432] env[61857]: INFO nova.compute.manager [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Took 8.17 seconds to spawn the instance on the hypervisor. [ 753.208475] env[61857]: DEBUG nova.compute.manager [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.209223] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe0fbef-187e-4851-9149-dd5d96bab4a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.238856] env[61857]: DEBUG nova.network.neutron [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Successfully updated port: af54a795-3d81-4edb-9d2c-05a68fc16a6f {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 753.248454] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 71920a62-090d-4df1-937e-87df3b043e28 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.257695] env[61857]: DEBUG oslo_vmware.api [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950790, 'name': PowerOffVM_Task, 'duration_secs': 0.26495} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.257955] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 753.258641] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 753.259618] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d96c556-2d72-47c6-b3a5-dabe8513c6a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.342368] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 753.342528] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 753.342712] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Deleting the datastore file [datastore2] c4f70321-f612-416e-b209-eb974dab9d49 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 753.342969] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7db54245-b554-4767-a367-e320917e0c11 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.353555] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529b5bbd-88e3-b2c7-4c00-e628f15286d2, 'name': SearchDatastore_Task, 'duration_secs': 0.010686} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.355975] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.355975] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] e286e6d0-25db-4d86-b9a2-3538bc3f624a/e286e6d0-25db-4d86-b9a2-3538bc3f624a.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 753.355975] env[61857]: DEBUG oslo_vmware.api [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for the task: (returnval){ [ 753.355975] env[61857]: value = "task-950793" [ 753.355975] env[61857]: _type = "Task" [ 753.355975] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.355975] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e6f2dc0-4ad2-4895-9e40-c1e7899027b2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.365578] env[61857]: DEBUG oslo_vmware.api [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.366758] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 753.366758] env[61857]: value = "task-950794" [ 753.366758] env[61857]: _type = "Task" [ 753.366758] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.373975] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.420363] env[61857]: DEBUG oslo_vmware.api [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950791, 'name': PowerOnVM_Task, 'duration_secs': 0.419966} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.420792] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 753.421035] env[61857]: INFO nova.compute.manager [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Took 5.82 seconds to spawn the instance on the hypervisor. [ 753.421399] env[61857]: DEBUG nova.compute.manager [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.422060] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb60d333-4422-4275-8f08-ce138eddbd59 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.649713] env[61857]: DEBUG nova.compute.manager [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.726864] env[61857]: INFO nova.compute.manager [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Took 27.72 seconds to build instance. [ 753.744438] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquiring lock "refresh_cache-7a800abe-ea0a-4077-a5da-dd60eab917e3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.744578] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquired lock "refresh_cache-7a800abe-ea0a-4077-a5da-dd60eab917e3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.744747] env[61857]: DEBUG nova.network.neutron [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.753438] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 28bf59ca-4ffe-4005-9a88-da0660ebb48a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.868992] env[61857]: DEBUG oslo_vmware.api [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.880882] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950794, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.939059] env[61857]: INFO nova.compute.manager [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Took 25.93 seconds to build instance. [ 754.172162] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.227894] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f82fb786-2071-4814-818a-737632c8b54a tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "d49008e6-7789-4785-8eea-9b3fa36c4ccb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.811s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.254108] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.277718] env[61857]: DEBUG nova.network.neutron [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.368843] env[61857]: DEBUG oslo_vmware.api [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Task: {'id': task-950793, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.532387} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.369299] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 754.369512] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 754.369704] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 754.369935] env[61857]: INFO nova.compute.manager [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Took 1.65 seconds to destroy the instance on the hypervisor. [ 754.370208] env[61857]: DEBUG oslo.service.loopingcall [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.373598] env[61857]: DEBUG nova.compute.manager [-] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.373720] env[61857]: DEBUG nova.network.neutron [-] [instance: c4f70321-f612-416e-b209-eb974dab9d49] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 754.381090] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950794, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.443759] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b0b49e4c-56ee-4ff6-a6cb-f8c0ad938a77 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "3cacef7b-dd24-4d33-9500-bbff03e342bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.508s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.451663] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "3fb5d24b-1767-43d9-bed5-833121962adb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.451982] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "3fb5d24b-1767-43d9-bed5-833121962adb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.453176] env[61857]: DEBUG nova.network.neutron [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Updating instance_info_cache with network_info: [{"id": "af54a795-3d81-4edb-9d2c-05a68fc16a6f", "address": "fa:16:3e:5b:4c:aa", "network": {"id": "6a0b4526-7fca-470f-95b8-55b31536fc0b", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1331352870-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba507a4694a3488bab7aa929657ba9f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf54a795-3d", "ovs_interfaceid": "af54a795-3d81-4edb-9d2c-05a68fc16a6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.730754] env[61857]: DEBUG nova.compute.manager [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 754.756833] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e91e99b9-1cd3-4345-af09-f14af4df1214 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.762412] env[61857]: DEBUG nova.compute.manager [req-27b2890c-5c4d-4033-b6c5-094706cc10c5 req-3184f21a-2f43-4a54-9ead-7889cb26f349 service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Received event network-vif-deleted-289c5ef4-b619-4720-bfc8-58a7b733beee {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 754.762412] env[61857]: INFO nova.compute.manager [req-27b2890c-5c4d-4033-b6c5-094706cc10c5 req-3184f21a-2f43-4a54-9ead-7889cb26f349 service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Neutron deleted interface 289c5ef4-b619-4720-bfc8-58a7b733beee; detaching it from the instance and deleting it from the info cache [ 754.762412] env[61857]: DEBUG nova.network.neutron [req-27b2890c-5c4d-4033-b6c5-094706cc10c5 req-3184f21a-2f43-4a54-9ead-7889cb26f349 service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.879137] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950794, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.947059] env[61857]: DEBUG nova.compute.manager [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 754.956613] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Releasing lock "refresh_cache-7a800abe-ea0a-4077-a5da-dd60eab917e3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.956917] env[61857]: DEBUG nova.compute.manager [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Instance network_info: |[{"id": "af54a795-3d81-4edb-9d2c-05a68fc16a6f", "address": "fa:16:3e:5b:4c:aa", "network": {"id": "6a0b4526-7fca-470f-95b8-55b31536fc0b", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1331352870-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba507a4694a3488bab7aa929657ba9f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf54a795-3d", "ovs_interfaceid": "af54a795-3d81-4edb-9d2c-05a68fc16a6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 754.957339] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:4c:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eed34ae1-5f7f-4deb-9db8-85eaa1e60c29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'af54a795-3d81-4edb-9d2c-05a68fc16a6f', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.965008] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Creating folder: Project (ba507a4694a3488bab7aa929657ba9f4). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 754.965304] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f8c51c1-2f4f-48c3-a37b-3e09b34350cb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.976020] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Created folder: Project (ba507a4694a3488bab7aa929657ba9f4) in parent group-v214027. [ 754.976222] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Creating folder: Instances. Parent ref: group-v214069. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 754.976479] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9353f8af-6f58-4fe1-8b95-3ce2f8db37d8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.984809] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Created folder: Instances in parent group-v214069. [ 754.985058] env[61857]: DEBUG oslo.service.loopingcall [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.985467] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 754.985700] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4d8dfba3-8ddd-418e-99c3-c7dac926916e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.006250] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 755.006250] env[61857]: value = "task-950797" [ 755.006250] env[61857]: _type = "Task" [ 755.006250] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.013588] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950797, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.092611] env[61857]: INFO nova.compute.manager [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Rebuilding instance [ 755.140018] env[61857]: DEBUG nova.compute.manager [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 755.140958] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f9406f-4f37-4f0c-95da-e9acb66116e1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.169612] env[61857]: DEBUG nova.network.neutron [-] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.211949] env[61857]: DEBUG nova.compute.manager [req-9bce5b14-7a34-4e58-9740-4bf5c0013323 req-152517fb-eb1e-427a-a571-96f2e3a772f7 service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Received event network-changed-af54a795-3d81-4edb-9d2c-05a68fc16a6f {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 755.212252] env[61857]: DEBUG nova.compute.manager [req-9bce5b14-7a34-4e58-9740-4bf5c0013323 req-152517fb-eb1e-427a-a571-96f2e3a772f7 service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Refreshing instance network info cache due to event network-changed-af54a795-3d81-4edb-9d2c-05a68fc16a6f. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 755.212252] env[61857]: DEBUG oslo_concurrency.lockutils [req-9bce5b14-7a34-4e58-9740-4bf5c0013323 req-152517fb-eb1e-427a-a571-96f2e3a772f7 service nova] Acquiring lock "refresh_cache-7a800abe-ea0a-4077-a5da-dd60eab917e3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.212420] env[61857]: DEBUG oslo_concurrency.lockutils [req-9bce5b14-7a34-4e58-9740-4bf5c0013323 req-152517fb-eb1e-427a-a571-96f2e3a772f7 service nova] Acquired lock "refresh_cache-7a800abe-ea0a-4077-a5da-dd60eab917e3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.212543] env[61857]: DEBUG nova.network.neutron [req-9bce5b14-7a34-4e58-9740-4bf5c0013323 req-152517fb-eb1e-427a-a571-96f2e3a772f7 service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Refreshing network info cache for port af54a795-3d81-4edb-9d2c-05a68fc16a6f {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 755.249395] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.264680] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 30d905ed-831d-44ae-807c-062de9a7e9fb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.266414] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-073ba47f-989c-4447-974d-112cff29eec5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.276057] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8033e3-f8a6-4c0c-a6e4-1eecab28b0aa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.303493] env[61857]: DEBUG nova.compute.manager [req-27b2890c-5c4d-4033-b6c5-094706cc10c5 req-3184f21a-2f43-4a54-9ead-7889cb26f349 service nova] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Detach interface failed, port_id=289c5ef4-b619-4720-bfc8-58a7b733beee, reason: Instance c4f70321-f612-416e-b209-eb974dab9d49 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 755.379553] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950794, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.825278} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.379817] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] e286e6d0-25db-4d86-b9a2-3538bc3f624a/e286e6d0-25db-4d86-b9a2-3538bc3f624a.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 755.380040] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 755.380387] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-664e3ede-6064-4727-9d7a-b2676507eadb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.386449] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 755.386449] env[61857]: value = "task-950798" [ 755.386449] env[61857]: _type = "Task" [ 755.386449] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.394742] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950798, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.465357] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.516576] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950797, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.652220] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 755.652652] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b6b0734-b5b2-4274-9607-ecac6da7ffea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.659629] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 755.659629] env[61857]: value = "task-950799" [ 755.659629] env[61857]: _type = "Task" [ 755.659629] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.668053] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950799, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.671875] env[61857]: INFO nova.compute.manager [-] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Took 1.30 seconds to deallocate network for instance. [ 755.769893] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 24d3d515-e03a-4b4e-bb8e-bc18537125ac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.899071] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950798, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067729} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.899358] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.900217] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8d0b24-5e84-47a6-a2df-a14ed89b81f8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.921338] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] e286e6d0-25db-4d86-b9a2-3538bc3f624a/e286e6d0-25db-4d86-b9a2-3538bc3f624a.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.921984] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45b5576d-6b71-4ad3-8674-ea3314babf9c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.942789] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 755.942789] env[61857]: value = "task-950800" [ 755.942789] env[61857]: _type = "Task" [ 755.942789] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.952468] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950800, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.016144] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950797, 'name': CreateVM_Task, 'duration_secs': 0.60855} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.016311] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 756.016999] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.017184] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.017494] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 756.017737] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12ccd631-d210-48aa-878a-3de1b6b2ebd0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.021945] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for the task: (returnval){ [ 756.021945] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525cea57-a9a6-b0b4-5814-9c79273739ce" [ 756.021945] env[61857]: _type = "Task" [ 756.021945] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.032071] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525cea57-a9a6-b0b4-5814-9c79273739ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.049106] env[61857]: DEBUG nova.network.neutron [req-9bce5b14-7a34-4e58-9740-4bf5c0013323 req-152517fb-eb1e-427a-a571-96f2e3a772f7 service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Updated VIF entry in instance network info cache for port af54a795-3d81-4edb-9d2c-05a68fc16a6f. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 756.049529] env[61857]: DEBUG nova.network.neutron [req-9bce5b14-7a34-4e58-9740-4bf5c0013323 req-152517fb-eb1e-427a-a571-96f2e3a772f7 service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Updating instance_info_cache with network_info: [{"id": "af54a795-3d81-4edb-9d2c-05a68fc16a6f", "address": "fa:16:3e:5b:4c:aa", "network": {"id": "6a0b4526-7fca-470f-95b8-55b31536fc0b", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1331352870-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba507a4694a3488bab7aa929657ba9f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf54a795-3d", "ovs_interfaceid": "af54a795-3d81-4edb-9d2c-05a68fc16a6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.171184] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950799, 'name': PowerOffVM_Task, 'duration_secs': 0.142723} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.171548] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 756.171775] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 756.172594] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da6fe83-6e3a-47e6-a726-6d6bf50aeb71 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.179794] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.180086] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 756.180329] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c29e16f-112c-49a0-8274-a8cbe653c57d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.202948] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 756.203177] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 756.203359] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Deleting the datastore file [datastore1] 3cacef7b-dd24-4d33-9500-bbff03e342bb {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 756.203612] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d578416e-8532-4d02-b577-cc8be63a1e00 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.209691] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 756.209691] env[61857]: value = "task-950802" [ 756.209691] env[61857]: _type = "Task" [ 756.209691] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.217267] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950802, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.272912] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 42d38f5e-628a-4030-85e1-3ec0595cf3c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.453453] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.532229] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525cea57-a9a6-b0b4-5814-9c79273739ce, 'name': SearchDatastore_Task, 'duration_secs': 0.009727} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.532583] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.532818] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 756.533069] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.533226] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.533411] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 756.533666] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff4f0a2f-722f-4e1a-91ee-5b23b3156407 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.543057] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 756.543057] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 756.543197] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc27bcf5-812b-4608-afdd-c12c54cab0da {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.549035] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for the task: (returnval){ [ 756.549035] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52bf8f36-92b8-e16e-d0fe-f8564d1607f3" [ 756.549035] env[61857]: _type = "Task" [ 756.549035] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.553720] env[61857]: DEBUG oslo_concurrency.lockutils [req-9bce5b14-7a34-4e58-9740-4bf5c0013323 req-152517fb-eb1e-427a-a571-96f2e3a772f7 service nova] Releasing lock "refresh_cache-7a800abe-ea0a-4077-a5da-dd60eab917e3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.556840] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52bf8f36-92b8-e16e-d0fe-f8564d1607f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.718888] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950802, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155298} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.719174] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 756.719352] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 756.719525] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 756.775903] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 6c4c1276-ae80-4f37-9ef9-4872dd415d24 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.954208] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950800, 'name': ReconfigVM_Task, 'duration_secs': 0.720755} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.954558] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Reconfigured VM instance instance-00000031 to attach disk [datastore1] e286e6d0-25db-4d86-b9a2-3538bc3f624a/e286e6d0-25db-4d86-b9a2-3538bc3f624a.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.955211] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-223d9037-2a88-437a-92c4-ada296016d07 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.965076] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 756.965076] env[61857]: value = "task-950803" [ 756.965076] env[61857]: _type = "Task" [ 756.965076] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.975143] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950803, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.060873] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52bf8f36-92b8-e16e-d0fe-f8564d1607f3, 'name': SearchDatastore_Task, 'duration_secs': 0.013243} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.060873] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08d70ffb-9620-4a07-a39d-a61ee8174f72 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.065490] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for the task: (returnval){ [ 757.065490] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]528803ab-5f1f-3715-5634-bc15e6a308e0" [ 757.065490] env[61857]: _type = "Task" [ 757.065490] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.072776] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528803ab-5f1f-3715-5634-bc15e6a308e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.280542] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 96ff157c-4471-4a12-ad99-0aafd6c2dce5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.475214] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950803, 'name': Rename_Task, 'duration_secs': 0.322421} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.475529] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 757.475771] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de89a423-b834-425c-9cb7-61f96c8d0764 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.483683] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 757.483683] env[61857]: value = "task-950804" [ 757.483683] env[61857]: _type = "Task" [ 757.483683] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.492488] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950804, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.576142] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528803ab-5f1f-3715-5634-bc15e6a308e0, 'name': SearchDatastore_Task, 'duration_secs': 0.010878} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.576459] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.576700] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 7a800abe-ea0a-4077-a5da-dd60eab917e3/7a800abe-ea0a-4077-a5da-dd60eab917e3.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 757.576958] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c71fa76f-28f7-4fcb-87d6-3c5845bb47ce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.583363] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for the task: (returnval){ [ 757.583363] env[61857]: value = "task-950805" [ 757.583363] env[61857]: _type = "Task" [ 757.583363] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.590994] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950805, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.754818] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.755212] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.755447] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.756016] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.756130] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.756296] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.756655] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.757010] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.757332] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.757579] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.757839] env[61857]: DEBUG nova.virt.hardware [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.758802] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c0bfaf-c64d-4a38-a5a5-3e84ed6aeaa9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.768123] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8708d612-7c8e-4d1b-b770-d07ec57cb30d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.785189] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 93e03e35-15a0-49e7-b1be-09178eabbfda has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.788016] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 757.795309] env[61857]: DEBUG oslo.service.loopingcall [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.796181] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 757.796495] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bdea3112-2340-4ad3-aa4f-ca16e672165d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.817932] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 757.817932] env[61857]: value = "task-950806" [ 757.817932] env[61857]: _type = "Task" [ 757.817932] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.830199] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950806, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.995161] env[61857]: DEBUG oslo_vmware.api [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950804, 'name': PowerOnVM_Task, 'duration_secs': 0.45767} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.995161] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 757.995415] env[61857]: INFO nova.compute.manager [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Took 7.84 seconds to spawn the instance on the hypervisor. [ 757.995614] env[61857]: DEBUG nova.compute.manager [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.996681] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3401b226-dc15-47fa-90de-616eb4c77010 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.094776] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950805, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452267} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.095063] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 7a800abe-ea0a-4077-a5da-dd60eab917e3/7a800abe-ea0a-4077-a5da-dd60eab917e3.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 758.095307] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 758.095598] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ebfe4f5d-649f-4c8e-a0a0-8dafe11f04f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.103683] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for the task: (returnval){ [ 758.103683] env[61857]: value = "task-950807" [ 758.103683] env[61857]: _type = "Task" [ 758.103683] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.113474] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950807, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.288098] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance bf2f46d4-5c6e-4493-b087-fe5a316b8c88 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.328082] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950806, 'name': CreateVM_Task, 'duration_secs': 0.378144} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.328270] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 758.328682] env[61857]: DEBUG oslo_concurrency.lockutils [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.328844] env[61857]: DEBUG oslo_concurrency.lockutils [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.329190] env[61857]: DEBUG oslo_concurrency.lockutils [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 758.329439] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60eb76f0-4f78-41ee-a905-ec843247d089 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.333993] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 758.333993] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52a19d28-eca6-10d8-6050-5f4b9448f380" [ 758.333993] env[61857]: _type = "Task" [ 758.333993] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.341600] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a19d28-eca6-10d8-6050-5f4b9448f380, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.514439] env[61857]: INFO nova.compute.manager [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Took 29.98 seconds to build instance. [ 758.613719] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950807, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06766} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.614385] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 758.614923] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f7d86d-1a40-4448-9593-ccb460ae5426 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.637023] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 7a800abe-ea0a-4077-a5da-dd60eab917e3/7a800abe-ea0a-4077-a5da-dd60eab917e3.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 758.637355] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b7a40d8-fe9c-41cd-aa5a-988f0c8f5aa0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.655683] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for the task: (returnval){ [ 758.655683] env[61857]: value = "task-950808" [ 758.655683] env[61857]: _type = "Task" [ 758.655683] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.663576] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950808, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.791310] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance a73f3163-b6fc-4764-b76a-451f823b5808 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.844638] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a19d28-eca6-10d8-6050-5f4b9448f380, 'name': SearchDatastore_Task, 'duration_secs': 0.040541} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.844936] env[61857]: DEBUG oslo_concurrency.lockutils [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.845187] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 758.845477] env[61857]: DEBUG oslo_concurrency.lockutils [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.845651] env[61857]: DEBUG oslo_concurrency.lockutils [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.845830] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 758.846090] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3be50c7f-75ca-473b-b8bb-dc6cdaee0f7f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.855385] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 758.855528] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 758.856211] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf2e9d8f-540f-4454-930d-3fd82b7fccc2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.861516] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 758.861516] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523d75b5-8d43-adc3-fc81-3c7e3b2aad41" [ 758.861516] env[61857]: _type = "Task" [ 758.861516] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.869203] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523d75b5-8d43-adc3-fc81-3c7e3b2aad41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.874693] env[61857]: INFO nova.compute.manager [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Rebuilding instance [ 758.916574] env[61857]: DEBUG nova.compute.manager [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 758.917467] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8b9fef-55f8-470a-b5d1-837111f87a7e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.016874] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f2a5a0c3-260f-4010-b48f-10456a0bf7ea tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "e286e6d0-25db-4d86-b9a2-3538bc3f624a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.987s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.165931] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950808, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.294818] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e5e64477-a1aa-4a4d-91a2-b17d912e09c3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.373505] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523d75b5-8d43-adc3-fc81-3c7e3b2aad41, 'name': SearchDatastore_Task, 'duration_secs': 0.01544} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.374269] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb5284cc-e29d-4b66-8ede-795e84ed268c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.379907] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 759.379907] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52c3f325-3c04-76f1-9c1b-22e2d1846513" [ 759.379907] env[61857]: _type = "Task" [ 759.379907] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.387467] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c3f325-3c04-76f1-9c1b-22e2d1846513, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.427554] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 759.427851] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4821aea-b8c7-418d-a836-4015bbfce24a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.435010] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 759.435010] env[61857]: value = "task-950809" [ 759.435010] env[61857]: _type = "Task" [ 759.435010] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.443513] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950809, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.519335] env[61857]: DEBUG nova.compute.manager [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.667159] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950808, 'name': ReconfigVM_Task, 'duration_secs': 0.713902} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.667466] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 7a800abe-ea0a-4077-a5da-dd60eab917e3/7a800abe-ea0a-4077-a5da-dd60eab917e3.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 759.668089] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c15249a-fb95-4521-8aa5-059328ff5d3d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.675483] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for the task: (returnval){ [ 759.675483] env[61857]: value = "task-950810" [ 759.675483] env[61857]: _type = "Task" [ 759.675483] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.683651] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950810, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.798174] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 16b447b5-426a-4478-9d44-ae32b41dee50 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.890389] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c3f325-3c04-76f1-9c1b-22e2d1846513, 'name': SearchDatastore_Task, 'duration_secs': 0.008668} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.890661] env[61857]: DEBUG oslo_concurrency.lockutils [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.890915] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 3cacef7b-dd24-4d33-9500-bbff03e342bb/3cacef7b-dd24-4d33-9500-bbff03e342bb.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 759.891190] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c21fb880-ec8b-4680-b2d3-b5287e1400b0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.899147] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 759.899147] env[61857]: value = "task-950811" [ 759.899147] env[61857]: _type = "Task" [ 759.899147] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.907314] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950811, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.944400] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950809, 'name': PowerOffVM_Task, 'duration_secs': 0.108124} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.944655] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 759.944866] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 759.945625] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3155f031-155f-4dc5-b4b0-0d3af590280a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.953672] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 759.954425] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b70458c-e924-4851-9f7d-957de4b9c4b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.976482] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 759.976673] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 759.976855] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Deleting the datastore file [datastore1] e286e6d0-25db-4d86-b9a2-3538bc3f624a {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 759.977142] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af1b25a9-7f66-4e90-967b-a9899d570a60 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.983536] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 759.983536] env[61857]: value = "task-950813" [ 759.983536] env[61857]: _type = "Task" [ 759.983536] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.991264] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950813, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.040801] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.186063] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950810, 'name': Rename_Task, 'duration_secs': 0.147338} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.186357] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 760.186620] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a47dedc0-7092-4e82-a4f8-17f58fdf04f2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.193375] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for the task: (returnval){ [ 760.193375] env[61857]: value = "task-950814" [ 760.193375] env[61857]: _type = "Task" [ 760.193375] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.201803] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950814, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.302037] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ece27044-3123-4685-a4ab-30eeb10337d1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.410135] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950811, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455625} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.410455] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 3cacef7b-dd24-4d33-9500-bbff03e342bb/3cacef7b-dd24-4d33-9500-bbff03e342bb.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 760.410665] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 760.410910] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-66d17752-1a03-486b-a40a-b342542c1fa9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.417198] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 760.417198] env[61857]: value = "task-950815" [ 760.417198] env[61857]: _type = "Task" [ 760.417198] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.425536] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950815, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.493744] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950813, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.120662} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.494016] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 760.494246] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 760.495154] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 760.704835] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950814, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.805631] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ad89cb5f-44d0-45c9-92a5-7f7ff3138e21 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.927506] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950815, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.200398} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.927893] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 760.928727] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f39012d-2770-40b4-b617-f4883e6f83eb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.948377] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 3cacef7b-dd24-4d33-9500-bbff03e342bb/3cacef7b-dd24-4d33-9500-bbff03e342bb.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 760.948981] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d8c56bd-b5bf-4be8-9afd-d207412efe5f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.967859] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 760.967859] env[61857]: value = "task-950816" [ 760.967859] env[61857]: _type = "Task" [ 760.967859] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.975865] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950816, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.204996] env[61857]: DEBUG oslo_vmware.api [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950814, 'name': PowerOnVM_Task, 'duration_secs': 0.765635} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.205294] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 761.205499] env[61857]: INFO nova.compute.manager [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Took 8.49 seconds to spawn the instance on the hypervisor. [ 761.205678] env[61857]: DEBUG nova.compute.manager [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 761.206452] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757c1c0f-5019-4160-ae5b-f5f661f89293 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.308702] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance d29f6e34-861d-4cbf-8cbd-ab193a55220b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.477708] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950816, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.539303] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.539546] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.539707] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.539948] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.540472] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.540724] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.540957] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.541145] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.541404] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.541539] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.541717] env[61857]: DEBUG nova.virt.hardware [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.542639] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f529531d-dc1c-45f4-82b6-34e3f87a5574 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.550893] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d8a10e-80ff-4111-b0ea-0caf2ad3e49c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.567578] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.574666] env[61857]: DEBUG oslo.service.loopingcall [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.575009] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 761.575248] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0602af80-00be-4f31-8d76-16a0cb4c8618 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.596403] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.596403] env[61857]: value = "task-950817" [ 761.596403] env[61857]: _type = "Task" [ 761.596403] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.605067] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950817, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.723150] env[61857]: INFO nova.compute.manager [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Took 32.16 seconds to build instance. [ 761.811944] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ddd4c436-405f-49f7-8c9b-de3b71725f63 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.982656] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950816, 'name': ReconfigVM_Task, 'duration_secs': 0.874024} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.984904] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 3cacef7b-dd24-4d33-9500-bbff03e342bb/3cacef7b-dd24-4d33-9500-bbff03e342bb.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.988602] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8a4f70e-8597-4f19-8426-c69e5c8877d0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.995790] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 761.995790] env[61857]: value = "task-950818" [ 761.995790] env[61857]: _type = "Task" [ 761.995790] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.004157] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950818, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.107419] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950817, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.225794] env[61857]: DEBUG oslo_concurrency.lockutils [None req-63a79c8d-95b0-415a-b5f9-c1c57108ba15 tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lock "7a800abe-ea0a-4077-a5da-dd60eab917e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.617s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.314930] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 7a9252da-2584-40ed-9d28-ca7341ed5165 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.508567] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950818, 'name': Rename_Task, 'duration_secs': 0.150774} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.508567] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 762.508567] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d43995e8-0c61-4876-8d86-530971d80a62 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.514540] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 762.514540] env[61857]: value = "task-950819" [ 762.514540] env[61857]: _type = "Task" [ 762.514540] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.523083] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950819, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.608910] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950817, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.729065] env[61857]: DEBUG nova.compute.manager [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.817804] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c27f6f5b-6423-4eee-be53-dc48c93df893 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.818211] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 762.818211] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 762.981554] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquiring lock "7a800abe-ea0a-4077-a5da-dd60eab917e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.981793] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lock "7a800abe-ea0a-4077-a5da-dd60eab917e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.981989] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquiring lock "7a800abe-ea0a-4077-a5da-dd60eab917e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.982189] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lock "7a800abe-ea0a-4077-a5da-dd60eab917e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.982359] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lock "7a800abe-ea0a-4077-a5da-dd60eab917e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.984804] env[61857]: INFO nova.compute.manager [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Terminating instance [ 762.986525] env[61857]: DEBUG nova.compute.manager [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.986716] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 762.987569] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328e7e9e-f804-483a-8a5a-d0f5c7c36b43 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.995723] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 762.995956] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4f25f82-a1e7-4077-917f-c330df358ff6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.001446] env[61857]: DEBUG oslo_vmware.api [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for the task: (returnval){ [ 763.001446] env[61857]: value = "task-950820" [ 763.001446] env[61857]: _type = "Task" [ 763.001446] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.012187] env[61857]: DEBUG oslo_vmware.api [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950820, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.029719] env[61857]: DEBUG oslo_vmware.api [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950819, 'name': PowerOnVM_Task, 'duration_secs': 0.423487} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.029719] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 763.029719] env[61857]: DEBUG nova.compute.manager [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 763.030445] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabf1210-65a9-4651-9054-e7467b85b39d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.112963] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950817, 'name': CreateVM_Task, 'duration_secs': 1.051397} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.112963] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 763.112963] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.112963] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.112963] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 763.112963] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f6e04b3-1d99-4c0e-8150-dc410b96c0b4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.125574] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 763.125574] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]522213dc-94f8-aae4-03f3-72e752452d92" [ 763.125574] env[61857]: _type = "Task" [ 763.125574] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.131782] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522213dc-94f8-aae4-03f3-72e752452d92, 'name': SearchDatastore_Task, 'duration_secs': 0.009997} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.135654] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.135654] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 763.135654] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.135654] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.135928] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 763.135928] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-405f2fa2-3e1e-49f4-99ed-f0027134c27d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.151081] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 763.154822] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 763.154822] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-babca9f3-eb09-47f7-adbe-9076db2b5ce0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.161064] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 763.161064] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]529fb15c-50ef-ad2f-d9cd-20a7b1f4afc5" [ 763.161064] env[61857]: _type = "Task" [ 763.161064] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.172639] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529fb15c-50ef-ad2f-d9cd-20a7b1f4afc5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.258796] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.306807] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9fa73c-eb8e-48cd-b209-0c4eb46c7406 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.314996] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34625a10-82a0-4904-a01c-460b590709b7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.348528] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e298cd23-6397-4c96-ace1-a673c58a79e1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.356102] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7742192-d684-4832-9b64-38e2fc081d85 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.371132] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.511724] env[61857]: DEBUG oslo_vmware.api [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950820, 'name': PowerOffVM_Task, 'duration_secs': 0.205004} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.511967] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 763.512154] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 763.512430] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d08f228c-ea52-48c7-8907-92a9b9e53365 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.548060] env[61857]: DEBUG oslo_concurrency.lockutils [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.594868] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 763.595118] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 763.595304] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Deleting the datastore file [datastore2] 7a800abe-ea0a-4077-a5da-dd60eab917e3 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 763.595573] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c5a91b2-0ace-4f7c-adff-cd26dd906b90 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.601995] env[61857]: DEBUG oslo_vmware.api [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for the task: (returnval){ [ 763.601995] env[61857]: value = "task-950822" [ 763.601995] env[61857]: _type = "Task" [ 763.601995] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.610298] env[61857]: DEBUG oslo_vmware.api [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950822, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.675711] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529fb15c-50ef-ad2f-d9cd-20a7b1f4afc5, 'name': SearchDatastore_Task, 'duration_secs': 0.009944} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.676513] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5a8ad0b-eec4-4cb0-a930-b47c79d64401 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.682185] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 763.682185] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52df6254-99db-c54e-dc54-b9724040e158" [ 763.682185] env[61857]: _type = "Task" [ 763.682185] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.689566] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52df6254-99db-c54e-dc54-b9724040e158, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.874158] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.029237] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "3cacef7b-dd24-4d33-9500-bbff03e342bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.031177] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "3cacef7b-dd24-4d33-9500-bbff03e342bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.031177] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "3cacef7b-dd24-4d33-9500-bbff03e342bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.031177] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "3cacef7b-dd24-4d33-9500-bbff03e342bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.031177] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "3cacef7b-dd24-4d33-9500-bbff03e342bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.032900] env[61857]: INFO nova.compute.manager [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Terminating instance [ 764.034634] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "refresh_cache-3cacef7b-dd24-4d33-9500-bbff03e342bb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.034781] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired lock "refresh_cache-3cacef7b-dd24-4d33-9500-bbff03e342bb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.034989] env[61857]: DEBUG nova.network.neutron [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.112093] env[61857]: DEBUG oslo_vmware.api [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Task: {'id': task-950822, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.463307} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.112355] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 764.112561] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 764.112749] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 764.112922] env[61857]: INFO nova.compute.manager [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 764.113198] env[61857]: DEBUG oslo.service.loopingcall [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.113400] env[61857]: DEBUG nova.compute.manager [-] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.113498] env[61857]: DEBUG nova.network.neutron [-] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 764.192731] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52df6254-99db-c54e-dc54-b9724040e158, 'name': SearchDatastore_Task, 'duration_secs': 0.03078} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.192994] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.193289] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] e286e6d0-25db-4d86-b9a2-3538bc3f624a/e286e6d0-25db-4d86-b9a2-3538bc3f624a.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 764.193547] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fed99460-c65d-4d74-84b9-05443cb68004 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.200479] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 764.200479] env[61857]: value = "task-950823" [ 764.200479] env[61857]: _type = "Task" [ 764.200479] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.208047] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950823, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.379473] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 764.379829] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.694s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.380229] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.654s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.382120] env[61857]: INFO nova.compute.claims [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.390672] env[61857]: DEBUG nova.compute.manager [req-302fb2d3-43c6-4dfb-b00c-b13fa3882390 req-d6ca3824-3efe-453d-9fe4-2139cfc995cd service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Received event network-vif-deleted-af54a795-3d81-4edb-9d2c-05a68fc16a6f {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 764.390858] env[61857]: INFO nova.compute.manager [req-302fb2d3-43c6-4dfb-b00c-b13fa3882390 req-d6ca3824-3efe-453d-9fe4-2139cfc995cd service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Neutron deleted interface af54a795-3d81-4edb-9d2c-05a68fc16a6f; detaching it from the instance and deleting it from the info cache [ 764.391243] env[61857]: DEBUG nova.network.neutron [req-302fb2d3-43c6-4dfb-b00c-b13fa3882390 req-d6ca3824-3efe-453d-9fe4-2139cfc995cd service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.552779] env[61857]: DEBUG nova.network.neutron [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.609385] env[61857]: DEBUG nova.network.neutron [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.711723] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950823, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.831848] env[61857]: DEBUG nova.network.neutron [-] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.893576] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be6eb871-bcb5-4570-8c9a-eb1653374be0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.905444] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba220f3d-3295-4dd2-87d1-c21c12ecaa73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.934767] env[61857]: DEBUG nova.compute.manager [req-302fb2d3-43c6-4dfb-b00c-b13fa3882390 req-d6ca3824-3efe-453d-9fe4-2139cfc995cd service nova] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Detach interface failed, port_id=af54a795-3d81-4edb-9d2c-05a68fc16a6f, reason: Instance 7a800abe-ea0a-4077-a5da-dd60eab917e3 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 765.113029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Releasing lock "refresh_cache-3cacef7b-dd24-4d33-9500-bbff03e342bb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.113127] env[61857]: DEBUG nova.compute.manager [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 765.113924] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 765.114836] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bb1633-a2e4-44cd-b9f3-9adb979698c3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.122529] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 765.122792] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec45cb20-92ea-4339-b9d1-f994527b93e1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.131625] env[61857]: DEBUG oslo_vmware.api [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 765.131625] env[61857]: value = "task-950824" [ 765.131625] env[61857]: _type = "Task" [ 765.131625] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.141213] env[61857]: DEBUG oslo_vmware.api [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950824, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.211176] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950823, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.929373} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.211471] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] e286e6d0-25db-4d86-b9a2-3538bc3f624a/e286e6d0-25db-4d86-b9a2-3538bc3f624a.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 765.211664] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 765.211944] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-353f4c19-4cfd-4202-bb5d-6ed7c8a1ca00 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.218355] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 765.218355] env[61857]: value = "task-950825" [ 765.218355] env[61857]: _type = "Task" [ 765.218355] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.230028] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950825, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.335060] env[61857]: INFO nova.compute.manager [-] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Took 1.22 seconds to deallocate network for instance. [ 765.641199] env[61857]: DEBUG oslo_vmware.api [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950824, 'name': PowerOffVM_Task, 'duration_secs': 0.11559} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.643465] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 765.643621] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 765.644042] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b59267eb-a5b6-4a15-8af3-789abf00fe1f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.670578] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 765.670793] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 765.670975] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Deleting the datastore file [datastore2] 3cacef7b-dd24-4d33-9500-bbff03e342bb {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 765.671243] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad5a9f4a-9749-45c7-8ea2-f7af1f227857 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.677980] env[61857]: DEBUG oslo_vmware.api [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 765.677980] env[61857]: value = "task-950827" [ 765.677980] env[61857]: _type = "Task" [ 765.677980] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.685761] env[61857]: DEBUG oslo_vmware.api [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950827, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.726223] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950825, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063082} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.728842] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 765.729820] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58a40c6-9697-4458-a6f9-5d967136908b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.749337] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] e286e6d0-25db-4d86-b9a2-3538bc3f624a/e286e6d0-25db-4d86-b9a2-3538bc3f624a.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 765.751705] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24a4d0d1-07f7-4864-8938-f21fdd9cb5ad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.771341] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 765.771341] env[61857]: value = "task-950828" [ 765.771341] env[61857]: _type = "Task" [ 765.771341] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.780421] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950828, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.787722] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-020fc3cc-2a65-41ce-90ac-8da58d7bd256 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.793800] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818388ff-fd01-4973-8cf0-c5e67f012c9d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.824749] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6dd5b1-0405-44a1-95d7-2ecd0b8ede8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.832519] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a41408-137f-469a-b7d3-cbb1e6c428ad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.846924] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.847577] env[61857]: DEBUG nova.compute.provider_tree [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.190032] env[61857]: DEBUG oslo_vmware.api [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950827, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095275} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.190369] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 766.190498] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 766.190729] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 766.190909] env[61857]: INFO nova.compute.manager [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Took 1.08 seconds to destroy the instance on the hypervisor. [ 766.191194] env[61857]: DEBUG oslo.service.loopingcall [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.191426] env[61857]: DEBUG nova.compute.manager [-] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.191533] env[61857]: DEBUG nova.network.neutron [-] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 766.207628] env[61857]: DEBUG nova.network.neutron [-] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.282661] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950828, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.350305] env[61857]: DEBUG nova.scheduler.client.report [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.711079] env[61857]: DEBUG nova.network.neutron [-] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.725485] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 766.782992] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950828, 'name': ReconfigVM_Task, 'duration_secs': 0.549159} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.782992] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Reconfigured VM instance instance-00000031 to attach disk [datastore2] e286e6d0-25db-4d86-b9a2-3538bc3f624a/e286e6d0-25db-4d86-b9a2-3538bc3f624a.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.782992] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53bd2d4d-b966-4628-b8ee-f7320075bb3e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.788550] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 766.788550] env[61857]: value = "task-950829" [ 766.788550] env[61857]: _type = "Task" [ 766.788550] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.795499] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950829, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.855864] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.856258] env[61857]: DEBUG nova.compute.manager [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.859063] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.098s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.860531] env[61857]: INFO nova.compute.claims [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.212832] env[61857]: INFO nova.compute.manager [-] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Took 1.02 seconds to deallocate network for instance. [ 767.234566] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.234856] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 767.234984] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Rebuilding the list of instances to heal {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 767.299540] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950829, 'name': Rename_Task, 'duration_secs': 0.136107} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.299786] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 767.300034] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91083245-95aa-42db-989f-bc223311dcee {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.306596] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 767.306596] env[61857]: value = "task-950830" [ 767.306596] env[61857]: _type = "Task" [ 767.306596] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.313995] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950830, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.364555] env[61857]: DEBUG nova.compute.utils [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.367487] env[61857]: DEBUG nova.compute.manager [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 767.367653] env[61857]: DEBUG nova.network.neutron [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 767.426340] env[61857]: DEBUG nova.policy [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3f6a9f3841f47cd999661673af1b7a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '02a79c0554e54bc3bbd5556b890e8f93', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.726068] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.740011] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Skipping network cache update for instance because it is being deleted. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 767.740355] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 767.740405] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 767.754597] env[61857]: DEBUG nova.network.neutron [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Successfully created port: e9861f7d-8566-414c-a73c-1e3caf722cd2 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.774420] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.774610] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquired lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.774766] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Forcefully refreshing network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 767.774952] env[61857]: DEBUG nova.objects.instance [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lazy-loading 'info_cache' on Instance uuid 2c187f64-8a36-4dfd-94e3-8ea944dbac24 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.817759] env[61857]: DEBUG oslo_vmware.api [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950830, 'name': PowerOnVM_Task, 'duration_secs': 0.401469} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.818164] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 767.818461] env[61857]: DEBUG nova.compute.manager [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.819286] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e8188d-e4fb-451a-9178-3e258aac78f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.871126] env[61857]: DEBUG nova.compute.manager [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 768.339150] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8945d1c1-7b1d-4256-8889-b3c107f56f9c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.341249] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.346863] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f700e94-4c13-4f76-8b2d-fc90b6bb9d84 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.380864] env[61857]: INFO nova.virt.block_device [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Booting with volume 3583f165-a169-43a4-975e-46e6fb77fbef at /dev/sda [ 768.383446] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51213f2e-a202-4297-8bfb-a51b2289c0d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.394700] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25dc59d3-5663-4e45-8e3f-314e9a14f5e2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.408628] env[61857]: DEBUG nova.compute.provider_tree [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.425923] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f00fbf3-f82e-4cd5-863a-b844ea3c7b95 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.437388] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66b697c-9b44-4928-89f4-95bd8754ddd0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.461620] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67aeb715-af98-4e6b-bce5-c4695d0d423f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.471119] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393fe403-793e-4c35-97dc-5d746bb2f04a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.496476] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee67fca-38b3-48a7-992b-7f68e8699353 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.503149] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bd702a-4dc8-469a-8a9b-007e35dc28a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.516061] env[61857]: DEBUG nova.virt.block_device [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Updating existing volume attachment record: c03f0f6e-084a-4ebb-9537-e11577676eac {{(pid=61857) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 768.912384] env[61857]: DEBUG nova.scheduler.client.report [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.918922] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "e286e6d0-25db-4d86-b9a2-3538bc3f624a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.918922] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "e286e6d0-25db-4d86-b9a2-3538bc3f624a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.918922] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "e286e6d0-25db-4d86-b9a2-3538bc3f624a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.918922] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "e286e6d0-25db-4d86-b9a2-3538bc3f624a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.919160] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "e286e6d0-25db-4d86-b9a2-3538bc3f624a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.920731] env[61857]: INFO nova.compute.manager [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Terminating instance [ 768.922312] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "refresh_cache-e286e6d0-25db-4d86-b9a2-3538bc3f624a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.922412] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquired lock "refresh_cache-e286e6d0-25db-4d86-b9a2-3538bc3f624a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.922539] env[61857]: DEBUG nova.network.neutron [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.415778] env[61857]: DEBUG nova.compute.manager [req-417d83d7-1e11-42ca-a68d-406570669c98 req-c1b1a9ef-70cd-45ea-b465-0c7246aa2ca8 service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Received event network-vif-plugged-e9861f7d-8566-414c-a73c-1e3caf722cd2 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 769.416012] env[61857]: DEBUG oslo_concurrency.lockutils [req-417d83d7-1e11-42ca-a68d-406570669c98 req-c1b1a9ef-70cd-45ea-b465-0c7246aa2ca8 service nova] Acquiring lock "c4883f84-e62b-4b59-9484-270d82dc34e0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.416213] env[61857]: DEBUG oslo_concurrency.lockutils [req-417d83d7-1e11-42ca-a68d-406570669c98 req-c1b1a9ef-70cd-45ea-b465-0c7246aa2ca8 service nova] Lock "c4883f84-e62b-4b59-9484-270d82dc34e0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.416382] env[61857]: DEBUG oslo_concurrency.lockutils [req-417d83d7-1e11-42ca-a68d-406570669c98 req-c1b1a9ef-70cd-45ea-b465-0c7246aa2ca8 service nova] Lock "c4883f84-e62b-4b59-9484-270d82dc34e0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.416544] env[61857]: DEBUG nova.compute.manager [req-417d83d7-1e11-42ca-a68d-406570669c98 req-c1b1a9ef-70cd-45ea-b465-0c7246aa2ca8 service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] No waiting events found dispatching network-vif-plugged-e9861f7d-8566-414c-a73c-1e3caf722cd2 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 769.416867] env[61857]: WARNING nova.compute.manager [req-417d83d7-1e11-42ca-a68d-406570669c98 req-c1b1a9ef-70cd-45ea-b465-0c7246aa2ca8 service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Received unexpected event network-vif-plugged-e9861f7d-8566-414c-a73c-1e3caf722cd2 for instance with vm_state building and task_state block_device_mapping. [ 769.417853] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.418375] env[61857]: DEBUG nova.compute.manager [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 769.422338] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.688s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.422338] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.422956] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.706s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.424418] env[61857]: INFO nova.compute.claims [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.455841] env[61857]: INFO nova.scheduler.client.report [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Deleted allocations for instance 6304714b-8988-4798-990f-d3e19774eb8f [ 769.458156] env[61857]: DEBUG nova.network.neutron [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.569057] env[61857]: DEBUG nova.network.neutron [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.572385] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Updating instance_info_cache with network_info: [{"id": "2104d242-18d6-4e99-b6d0-1dd9efc2778b", "address": "fa:16:3e:95:af:9d", "network": {"id": "ca954f70-9b06-47d1-8347-d99e04452f69", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1065246308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6be7236fbc844deb8011502c4910401f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2104d242-18", "ovs_interfaceid": "2104d242-18d6-4e99-b6d0-1dd9efc2778b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.621975] env[61857]: DEBUG nova.network.neutron [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Successfully updated port: e9861f7d-8566-414c-a73c-1e3caf722cd2 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 769.934049] env[61857]: DEBUG nova.compute.utils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.936013] env[61857]: DEBUG nova.compute.manager [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 769.936013] env[61857]: DEBUG nova.network.neutron [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 769.968123] env[61857]: DEBUG oslo_concurrency.lockutils [None req-72455026-ab14-4ebd-82de-83b045becdd8 tempest-ServersAaction247Test-332744453 tempest-ServersAaction247Test-332744453-project-member] Lock "6304714b-8988-4798-990f-d3e19774eb8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.985s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.004640] env[61857]: DEBUG nova.policy [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd5c2aa088da5442b9332221da335f2d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4725d163dec74a53940dc24ae88927f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 770.072587] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Releasing lock "refresh_cache-e286e6d0-25db-4d86-b9a2-3538bc3f624a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.073305] env[61857]: DEBUG nova.compute.manager [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 770.073305] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 770.074451] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a02cdb-30e7-4cba-949b-9a9c70dfc0a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.077976] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Releasing lock "refresh_cache-2c187f64-8a36-4dfd-94e3-8ea944dbac24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.078233] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Updated the network info_cache for instance {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 770.078436] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.078596] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.078750] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.078897] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.079073] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.079317] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 770.082165] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 770.082408] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dff93b8b-5582-44e7-a8e3-4259566897f0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.088495] env[61857]: DEBUG oslo_vmware.api [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 770.088495] env[61857]: value = "task-950831" [ 770.088495] env[61857]: _type = "Task" [ 770.088495] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.096700] env[61857]: DEBUG oslo_vmware.api [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.124191] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Acquiring lock "refresh_cache-c4883f84-e62b-4b59-9484-270d82dc34e0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.124191] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Acquired lock "refresh_cache-c4883f84-e62b-4b59-9484-270d82dc34e0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.124191] env[61857]: DEBUG nova.network.neutron [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 770.332228] env[61857]: DEBUG nova.network.neutron [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Successfully created port: 5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.441419] env[61857]: DEBUG nova.compute.manager [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 770.603965] env[61857]: DEBUG oslo_vmware.api [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950831, 'name': PowerOffVM_Task, 'duration_secs': 0.192257} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.603965] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 770.603965] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 770.604389] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18296a77-1ea5-4369-a7d4-399379dc0f7d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.613530] env[61857]: DEBUG nova.compute.manager [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.614494] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.614494] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.614494] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.614660] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.614841] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.615015] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.615232] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.615396] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.615591] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.615914] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.616151] env[61857]: DEBUG nova.virt.hardware [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.620254] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a0a6255-945e-47fd-b5f3-61811450797e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.631601] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44309967-c892-47c9-9508-64a83bb7fb03 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.636409] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 770.636409] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 770.636580] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Deleting the datastore file [datastore2] e286e6d0-25db-4d86-b9a2-3538bc3f624a {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 770.639430] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9cd506d9-d523-4d9a-93e8-4b460799b2d8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.655957] env[61857]: DEBUG oslo_vmware.api [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for the task: (returnval){ [ 770.655957] env[61857]: value = "task-950833" [ 770.655957] env[61857]: _type = "Task" [ 770.655957] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.664520] env[61857]: DEBUG oslo_vmware.api [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.673273] env[61857]: DEBUG nova.network.neutron [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.850679] env[61857]: DEBUG nova.network.neutron [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Updating instance_info_cache with network_info: [{"id": "e9861f7d-8566-414c-a73c-1e3caf722cd2", "address": "fa:16:3e:4d:95:17", "network": {"id": "3a7e17fb-072f-4bb8-87df-b69f668b4ee0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-188702091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "02a79c0554e54bc3bbd5556b890e8f93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e1a5c1-4ae7-409b-8de7-d401684ef60d", "external-id": "nsx-vlan-transportzone-740", "segmentation_id": 740, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9861f7d-85", "ovs_interfaceid": "e9861f7d-8566-414c-a73c-1e3caf722cd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.918921] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f8ed15-e899-415e-b7db-cc6f101565b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.926194] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4f4699-255e-440e-a349-0c77e2f6c697 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.959297] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d110da-8aaf-4624-b11d-31ca436c754e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.966449] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a2c8e4-32f0-41c4-88f0-cec93110b7e5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.979124] env[61857]: DEBUG nova.compute.provider_tree [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.168210] env[61857]: DEBUG oslo_vmware.api [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Task: {'id': task-950833, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093562} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.168496] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 771.168679] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 771.168853] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 771.169035] env[61857]: INFO nova.compute.manager [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Took 1.10 seconds to destroy the instance on the hypervisor. [ 771.169288] env[61857]: DEBUG oslo.service.loopingcall [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.169478] env[61857]: DEBUG nova.compute.manager [-] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.169570] env[61857]: DEBUG nova.network.neutron [-] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 771.185031] env[61857]: DEBUG nova.network.neutron [-] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 771.353407] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Releasing lock "refresh_cache-c4883f84-e62b-4b59-9484-270d82dc34e0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.353759] env[61857]: DEBUG nova.compute.manager [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Instance network_info: |[{"id": "e9861f7d-8566-414c-a73c-1e3caf722cd2", "address": "fa:16:3e:4d:95:17", "network": {"id": "3a7e17fb-072f-4bb8-87df-b69f668b4ee0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-188702091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "02a79c0554e54bc3bbd5556b890e8f93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e1a5c1-4ae7-409b-8de7-d401684ef60d", "external-id": "nsx-vlan-transportzone-740", "segmentation_id": 740, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9861f7d-85", "ovs_interfaceid": "e9861f7d-8566-414c-a73c-1e3caf722cd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 771.354423] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:95:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69e1a5c1-4ae7-409b-8de7-d401684ef60d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e9861f7d-8566-414c-a73c-1e3caf722cd2', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.361588] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Creating folder: Project (02a79c0554e54bc3bbd5556b890e8f93). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 771.361860] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d87a48b-3b4c-4859-93af-1e7747395b27 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.375430] env[61857]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 771.375606] env[61857]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61857) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 771.375933] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Folder already exists: Project (02a79c0554e54bc3bbd5556b890e8f93). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 771.376146] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Creating folder: Instances. Parent ref: group-v214040. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 771.376383] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4a1a5441-9990-418f-8813-06b64c8042fb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.385751] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Created folder: Instances in parent group-v214040. [ 771.385977] env[61857]: DEBUG oslo.service.loopingcall [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.386193] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 771.386396] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c80e476f-ab67-496d-9404-2a18ce25ed6a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.405154] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.405154] env[61857]: value = "task-950836" [ 771.405154] env[61857]: _type = "Task" [ 771.405154] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.412700] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950836, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.440010] env[61857]: DEBUG nova.compute.manager [req-bf26a351-1cdb-4291-8dcf-b847bdf900b8 req-48bb7065-5e09-4431-87ed-a3e21c035e2d service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Received event network-changed-e9861f7d-8566-414c-a73c-1e3caf722cd2 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 771.440204] env[61857]: DEBUG nova.compute.manager [req-bf26a351-1cdb-4291-8dcf-b847bdf900b8 req-48bb7065-5e09-4431-87ed-a3e21c035e2d service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Refreshing instance network info cache due to event network-changed-e9861f7d-8566-414c-a73c-1e3caf722cd2. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 771.440446] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf26a351-1cdb-4291-8dcf-b847bdf900b8 req-48bb7065-5e09-4431-87ed-a3e21c035e2d service nova] Acquiring lock "refresh_cache-c4883f84-e62b-4b59-9484-270d82dc34e0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.440593] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf26a351-1cdb-4291-8dcf-b847bdf900b8 req-48bb7065-5e09-4431-87ed-a3e21c035e2d service nova] Acquired lock "refresh_cache-c4883f84-e62b-4b59-9484-270d82dc34e0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.440754] env[61857]: DEBUG nova.network.neutron [req-bf26a351-1cdb-4291-8dcf-b847bdf900b8 req-48bb7065-5e09-4431-87ed-a3e21c035e2d service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Refreshing network info cache for port e9861f7d-8566-414c-a73c-1e3caf722cd2 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 771.463396] env[61857]: DEBUG nova.compute.manager [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 771.482793] env[61857]: DEBUG nova.scheduler.client.report [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.492115] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.492377] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.492539] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.492721] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.492866] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.493020] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.493227] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.493385] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.493548] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.493706] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.493884] env[61857]: DEBUG nova.virt.hardware [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.494743] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d4b9a9-ec0d-4f8a-8da8-ff8c5026287b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.503380] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5fd84d4-745d-486a-b299-cb68cecf9bf8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.687684] env[61857]: DEBUG nova.network.neutron [-] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.904864] env[61857]: DEBUG nova.network.neutron [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Successfully updated port: 5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 771.918494] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950836, 'name': CreateVM_Task, 'duration_secs': 0.278862} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.922833] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 771.922833] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214043', 'volume_id': '3583f165-a169-43a4-975e-46e6fb77fbef', 'name': 'volume-3583f165-a169-43a4-975e-46e6fb77fbef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4883f84-e62b-4b59-9484-270d82dc34e0', 'attached_at': '', 'detached_at': '', 'volume_id': '3583f165-a169-43a4-975e-46e6fb77fbef', 'serial': '3583f165-a169-43a4-975e-46e6fb77fbef'}, 'mount_device': '/dev/sda', 'guest_format': None, 'device_type': None, 'attachment_id': 'c03f0f6e-084a-4ebb-9537-e11577676eac', 'delete_on_termination': True, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=61857) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 771.922833] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Root volume attach. Driver type: vmdk {{(pid=61857) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 771.923163] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015752ae-3f93-4920-9777-3e91cac6689a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.928785] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e19651-a31c-46f8-abc8-34f26c31eaf9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.936068] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25c436d-75b4-47f3-9cc8-152662035d2b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.944435] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-86e2f9aa-4f7d-4042-8e2c-142c8c6ee09a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.951921] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for the task: (returnval){ [ 771.951921] env[61857]: value = "task-950837" [ 771.951921] env[61857]: _type = "Task" [ 771.951921] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.961268] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950837, 'name': RelocateVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.988770] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.989343] env[61857]: DEBUG nova.compute.manager [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.995871] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.739s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.995871] env[61857]: INFO nova.compute.claims [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.186050] env[61857]: DEBUG nova.network.neutron [req-bf26a351-1cdb-4291-8dcf-b847bdf900b8 req-48bb7065-5e09-4431-87ed-a3e21c035e2d service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Updated VIF entry in instance network info cache for port e9861f7d-8566-414c-a73c-1e3caf722cd2. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 772.186050] env[61857]: DEBUG nova.network.neutron [req-bf26a351-1cdb-4291-8dcf-b847bdf900b8 req-48bb7065-5e09-4431-87ed-a3e21c035e2d service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Updating instance_info_cache with network_info: [{"id": "e9861f7d-8566-414c-a73c-1e3caf722cd2", "address": "fa:16:3e:4d:95:17", "network": {"id": "3a7e17fb-072f-4bb8-87df-b69f668b4ee0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-188702091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "02a79c0554e54bc3bbd5556b890e8f93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e1a5c1-4ae7-409b-8de7-d401684ef60d", "external-id": "nsx-vlan-transportzone-740", "segmentation_id": 740, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9861f7d-85", "ovs_interfaceid": "e9861f7d-8566-414c-a73c-1e3caf722cd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.194021] env[61857]: INFO nova.compute.manager [-] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Took 1.02 seconds to deallocate network for instance. [ 772.410308] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.410308] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.410538] env[61857]: DEBUG nova.network.neutron [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 772.462369] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950837, 'name': RelocateVM_Task, 'duration_secs': 0.347368} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.462369] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Volume attach. Driver type: vmdk {{(pid=61857) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 772.462566] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214043', 'volume_id': '3583f165-a169-43a4-975e-46e6fb77fbef', 'name': 'volume-3583f165-a169-43a4-975e-46e6fb77fbef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4883f84-e62b-4b59-9484-270d82dc34e0', 'attached_at': '', 'detached_at': '', 'volume_id': '3583f165-a169-43a4-975e-46e6fb77fbef', 'serial': '3583f165-a169-43a4-975e-46e6fb77fbef'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 772.463798] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666595ff-af68-4a90-8aba-02066c518360 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.479358] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a68d3a9-d114-4708-aae2-f32d8b9819b7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.502703] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] volume-3583f165-a169-43a4-975e-46e6fb77fbef/volume-3583f165-a169-43a4-975e-46e6fb77fbef.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.506747] env[61857]: DEBUG nova.compute.utils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.507922] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cdc1ebc-6ab9-4931-aa7a-378cfc1f8537 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.527050] env[61857]: DEBUG nova.compute.manager [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 772.527158] env[61857]: DEBUG nova.network.neutron [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 772.536123] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for the task: (returnval){ [ 772.536123] env[61857]: value = "task-950838" [ 772.536123] env[61857]: _type = "Task" [ 772.536123] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.544641] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950838, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.595073] env[61857]: DEBUG nova.policy [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f719cbf1755d4c34bef92dab6679cb56', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ef111c8b3ff4f2383b4e81db026792b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.687909] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf26a351-1cdb-4291-8dcf-b847bdf900b8 req-48bb7065-5e09-4431-87ed-a3e21c035e2d service nova] Releasing lock "refresh_cache-c4883f84-e62b-4b59-9484-270d82dc34e0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.696556] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.921311] env[61857]: DEBUG nova.network.neutron [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Successfully created port: 90c3f890-dd3a-4d33-92c0-a6ec8632f14b {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.965858] env[61857]: DEBUG nova.network.neutron [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.982155] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2f0c1b-ca2c-4b9a-ac4b-04cc25306627 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.992510] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0168ae4-42a7-42d8-beff-56cfa01db6a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.030925] env[61857]: DEBUG nova.compute.manager [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 773.037668] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e685ab-c41d-4133-b00c-1b3fb22685f9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.051815] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b70bcb-e3f8-480b-a163-8e0006dc2619 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.058444] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950838, 'name': ReconfigVM_Task, 'duration_secs': 0.247821} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.058444] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Reconfigured VM instance instance-00000033 to attach disk [datastore2] volume-3583f165-a169-43a4-975e-46e6fb77fbef/volume-3583f165-a169-43a4-975e-46e6fb77fbef.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.061056] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-264c31ea-0be2-4729-a937-14078a8301c3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.080511] env[61857]: DEBUG nova.compute.provider_tree [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.093023] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for the task: (returnval){ [ 773.093023] env[61857]: value = "task-950839" [ 773.093023] env[61857]: _type = "Task" [ 773.093023] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.099476] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950839, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.200637] env[61857]: DEBUG nova.network.neutron [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updating instance_info_cache with network_info: [{"id": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "address": "fa:16:3e:dc:03:c9", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5da8191c-5f", "ovs_interfaceid": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.484964] env[61857]: DEBUG nova.compute.manager [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Received event network-vif-plugged-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 773.484964] env[61857]: DEBUG oslo_concurrency.lockutils [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] Acquiring lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.484964] env[61857]: DEBUG oslo_concurrency.lockutils [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.484964] env[61857]: DEBUG oslo_concurrency.lockutils [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.484964] env[61857]: DEBUG nova.compute.manager [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] No waiting events found dispatching network-vif-plugged-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 773.485494] env[61857]: WARNING nova.compute.manager [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Received unexpected event network-vif-plugged-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 for instance with vm_state building and task_state spawning. [ 773.485494] env[61857]: DEBUG nova.compute.manager [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Received event network-changed-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 773.485494] env[61857]: DEBUG nova.compute.manager [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Refreshing instance network info cache due to event network-changed-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 773.485494] env[61857]: DEBUG oslo_concurrency.lockutils [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] Acquiring lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.585802] env[61857]: DEBUG nova.scheduler.client.report [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.603727] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950839, 'name': ReconfigVM_Task, 'duration_secs': 0.119161} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.604040] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214043', 'volume_id': '3583f165-a169-43a4-975e-46e6fb77fbef', 'name': 'volume-3583f165-a169-43a4-975e-46e6fb77fbef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4883f84-e62b-4b59-9484-270d82dc34e0', 'attached_at': '', 'detached_at': '', 'volume_id': '3583f165-a169-43a4-975e-46e6fb77fbef', 'serial': '3583f165-a169-43a4-975e-46e6fb77fbef'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 773.604565] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d20fd43-243f-4737-ba3a-badc151fbae6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.611537] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for the task: (returnval){ [ 773.611537] env[61857]: value = "task-950840" [ 773.611537] env[61857]: _type = "Task" [ 773.611537] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.619128] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950840, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.703604] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.704073] env[61857]: DEBUG nova.compute.manager [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Instance network_info: |[{"id": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "address": "fa:16:3e:dc:03:c9", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5da8191c-5f", "ovs_interfaceid": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 773.704406] env[61857]: DEBUG oslo_concurrency.lockutils [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] Acquired lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.704608] env[61857]: DEBUG nova.network.neutron [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Refreshing network info cache for port 5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.706102] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:03:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 773.714091] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Creating folder: Project (4725d163dec74a53940dc24ae88927f8). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 773.715156] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52cb8f6d-499b-48f1-918b-c6ae94890ff7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.726092] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Created folder: Project (4725d163dec74a53940dc24ae88927f8) in parent group-v214027. [ 773.726337] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Creating folder: Instances. Parent ref: group-v214076. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 773.726608] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a8ac9e9d-0c51-4430-9b8c-5461ea508aba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.735990] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Created folder: Instances in parent group-v214076. [ 773.736571] env[61857]: DEBUG oslo.service.loopingcall [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.736571] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 773.736756] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28402929-6da8-4595-8ece-27c65fa8ab41 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.756391] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 773.756391] env[61857]: value = "task-950843" [ 773.756391] env[61857]: _type = "Task" [ 773.756391] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.763788] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950843, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.788321] env[61857]: DEBUG nova.compute.manager [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 774.791096] env[61857]: DEBUG nova.network.neutron [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Successfully updated port: 90c3f890-dd3a-4d33-92c0-a6ec8632f14b {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 774.792568] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.800s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.793030] env[61857]: DEBUG nova.compute.manager [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.798697] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.867s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.800043] env[61857]: INFO nova.compute.claims [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.823015] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950843, 'name': CreateVM_Task, 'duration_secs': 0.561506} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.823015] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950840, 'name': Rename_Task, 'duration_secs': 0.13136} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.823015] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 774.823015] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 774.823015] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.823527] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.823527] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 774.823527] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-37e6c591-87e1-4a89-9387-9a8db4196c07 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.823527] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28156a3e-e80e-43bd-b45a-6e7b0707d7bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.827830] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 774.827830] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52af6a13-1b63-de07-fede-d50a20957667" [ 774.827830] env[61857]: _type = "Task" [ 774.827830] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.833860] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.834131] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.834313] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.834526] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.834739] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.834851] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.835071] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.835231] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.835400] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.835561] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.836020] env[61857]: DEBUG nova.virt.hardware [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.836090] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for the task: (returnval){ [ 774.836090] env[61857]: value = "task-950844" [ 774.836090] env[61857]: _type = "Task" [ 774.836090] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.836718] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d2325b-1249-40d5-b156-402c5ad1fbc6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.846434] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52af6a13-1b63-de07-fede-d50a20957667, 'name': SearchDatastore_Task, 'duration_secs': 0.009866} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.850614] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.850847] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 774.851137] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.851297] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.851477] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 774.855209] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-882c674a-5069-4510-be75-9512d87c40d4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.858097] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b4f947-d7f9-4661-bcdd-612c0410f3a7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.862137] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950844, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.873422] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 774.873624] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 774.875288] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cf90942-b915-4368-a245-b0dfbce5440a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.882550] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 774.882550] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52d705ac-162a-616b-1935-d7e09454d494" [ 774.882550] env[61857]: _type = "Task" [ 774.882550] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.889041] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52d705ac-162a-616b-1935-d7e09454d494, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.051439] env[61857]: DEBUG nova.network.neutron [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updated VIF entry in instance network info cache for port 5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 775.051803] env[61857]: DEBUG nova.network.neutron [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updating instance_info_cache with network_info: [{"id": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "address": "fa:16:3e:dc:03:c9", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5da8191c-5f", "ovs_interfaceid": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.301649] env[61857]: DEBUG nova.compute.utils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.302237] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.302938] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.303147] env[61857]: DEBUG nova.network.neutron [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 775.304439] env[61857]: DEBUG nova.compute.manager [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 775.304946] env[61857]: DEBUG nova.network.neutron [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 775.349040] env[61857]: DEBUG oslo_vmware.api [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950844, 'name': PowerOnVM_Task, 'duration_secs': 0.496248} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.349326] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 775.349550] env[61857]: INFO nova.compute.manager [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Took 4.74 seconds to spawn the instance on the hypervisor. [ 775.350813] env[61857]: DEBUG nova.compute.manager [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 775.351215] env[61857]: DEBUG nova.policy [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a00d6256b5f4f78bc0fd160605123bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3eca39f9d1f244a6b73b33326f4eea7c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.353103] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e970fad-f027-47f2-952a-18d969f341e2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.392146] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52d705ac-162a-616b-1935-d7e09454d494, 'name': SearchDatastore_Task, 'duration_secs': 0.010134} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.392819] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec9fbda7-d2d7-4c09-bb6f-594d8139845f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.398405] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 775.398405] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52c8ba33-e142-8223-8886-8fb3c1e96c8c" [ 775.398405] env[61857]: _type = "Task" [ 775.398405] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.405934] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c8ba33-e142-8223-8886-8fb3c1e96c8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.490707] env[61857]: DEBUG nova.compute.manager [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Received event network-vif-plugged-90c3f890-dd3a-4d33-92c0-a6ec8632f14b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 775.490707] env[61857]: DEBUG oslo_concurrency.lockutils [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] Acquiring lock "71920a62-090d-4df1-937e-87df3b043e28-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.490707] env[61857]: DEBUG oslo_concurrency.lockutils [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] Lock "71920a62-090d-4df1-937e-87df3b043e28-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.490707] env[61857]: DEBUG oslo_concurrency.lockutils [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] Lock "71920a62-090d-4df1-937e-87df3b043e28-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.490707] env[61857]: DEBUG nova.compute.manager [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] No waiting events found dispatching network-vif-plugged-90c3f890-dd3a-4d33-92c0-a6ec8632f14b {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 775.491252] env[61857]: WARNING nova.compute.manager [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Received unexpected event network-vif-plugged-90c3f890-dd3a-4d33-92c0-a6ec8632f14b for instance with vm_state building and task_state spawning. [ 775.491252] env[61857]: DEBUG nova.compute.manager [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Received event network-changed-90c3f890-dd3a-4d33-92c0-a6ec8632f14b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 775.491252] env[61857]: DEBUG nova.compute.manager [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Refreshing instance network info cache due to event network-changed-90c3f890-dd3a-4d33-92c0-a6ec8632f14b. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 775.491252] env[61857]: DEBUG oslo_concurrency.lockutils [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] Acquiring lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.554177] env[61857]: DEBUG oslo_concurrency.lockutils [req-95ef4769-9a91-47f2-aba7-4837ba0cf2e5 req-fe80d4a1-2db6-45cf-873a-4f9153ce4379 service nova] Releasing lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.628907] env[61857]: DEBUG nova.network.neutron [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Successfully created port: 45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.689422] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.689632] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Cleaning up deleted instances {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 775.806745] env[61857]: DEBUG nova.compute.manager [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 775.873770] env[61857]: INFO nova.compute.manager [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Took 41.17 seconds to build instance. [ 775.877508] env[61857]: DEBUG nova.network.neutron [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.911473] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c8ba33-e142-8223-8886-8fb3c1e96c8c, 'name': SearchDatastore_Task, 'duration_secs': 0.009245} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.916045] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.916335] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 781a2790-e317-46fe-9be8-40c9e5f1f771/781a2790-e317-46fe-9be8-40c9e5f1f771.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 775.916827] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-077aef66-3302-4c2a-afc9-41daf4148414 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.926356] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 775.926356] env[61857]: value = "task-950845" [ 775.926356] env[61857]: _type = "Task" [ 775.926356] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.937120] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.113140] env[61857]: DEBUG nova.network.neutron [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance_info_cache with network_info: [{"id": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "address": "fa:16:3e:32:96:34", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c3f890-dd", "ovs_interfaceid": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.194718] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] There are 3 instances to clean {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 776.195014] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 6304714b-8988-4798-990f-d3e19774eb8f] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 776.310015] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bce56f8-a748-495e-bb19-aefd8efe15c3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.321878] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ed7104-9d67-42e3-98c4-adf349a54fb9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.359355] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e5d482-b18b-47a1-9724-07e53f9a5c8d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.367456] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c764784-94e7-4521-945d-1a751b5ebd0e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.381896] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37eae172-e37c-4010-bba8-f1f40ecab8e6 tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Lock "c4883f84-e62b-4b59-9484-270d82dc34e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.171s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.382625] env[61857]: DEBUG nova.compute.provider_tree [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.436039] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950845, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446435} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.436314] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 781a2790-e317-46fe-9be8-40c9e5f1f771/781a2790-e317-46fe-9be8-40c9e5f1f771.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 776.436522] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 776.436766] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba210dba-79b6-42eb-9e51-0de6b3dc4aa4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.442740] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 776.442740] env[61857]: value = "task-950846" [ 776.442740] env[61857]: _type = "Task" [ 776.442740] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.450659] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950846, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.616063] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.616528] env[61857]: DEBUG nova.compute.manager [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Instance network_info: |[{"id": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "address": "fa:16:3e:32:96:34", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c3f890-dd", "ovs_interfaceid": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 776.617022] env[61857]: DEBUG oslo_concurrency.lockutils [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] Acquired lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.617291] env[61857]: DEBUG nova.network.neutron [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Refreshing network info cache for port 90c3f890-dd3a-4d33-92c0-a6ec8632f14b {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 776.618526] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:96:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90c3f890-dd3a-4d33-92c0-a6ec8632f14b', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 776.626364] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Creating folder: Project (5ef111c8b3ff4f2383b4e81db026792b). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 776.627273] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05c8f3a2-70e9-4f9d-878b-e4c21b68830e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.641860] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Created folder: Project (5ef111c8b3ff4f2383b4e81db026792b) in parent group-v214027. [ 776.641860] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Creating folder: Instances. Parent ref: group-v214079. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 776.641860] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15e68156-8587-4cf9-9eae-3623fe6ad916 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.652431] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Created folder: Instances in parent group-v214079. [ 776.653346] env[61857]: DEBUG oslo.service.loopingcall [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.653560] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 776.654377] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74f91ae5-bf18-415d-a6ed-60c03e5d496d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.673886] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 776.673886] env[61857]: value = "task-950849" [ 776.673886] env[61857]: _type = "Task" [ 776.673886] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.681406] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950849, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.702213] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 2e7c154b-dab2-4384-bdb8-d098967654c5] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 776.818289] env[61857]: DEBUG nova.compute.manager [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.846434] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.846721] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.846882] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.847072] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.847225] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.847387] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.847596] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.847755] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.847922] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.848114] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.848365] env[61857]: DEBUG nova.virt.hardware [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.849231] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d3f871-8e4b-4db7-8776-5b6eaa32eb3a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.857433] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee22ef65-a3ed-465c-b196-9ea6ad29daba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.885990] env[61857]: DEBUG nova.scheduler.client.report [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.889427] env[61857]: DEBUG nova.compute.manager [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 776.952953] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950846, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064997} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.953250] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 776.954082] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3001e89-bce0-4bc5-bcb9-2924b14e39f3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.977243] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 781a2790-e317-46fe-9be8-40c9e5f1f771/781a2790-e317-46fe-9be8-40c9e5f1f771.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 776.977851] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9890dafa-2f2e-475c-bce9-a091918718ef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.998761] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 776.998761] env[61857]: value = "task-950850" [ 776.998761] env[61857]: _type = "Task" [ 776.998761] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.006765] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950850, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.184293] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950849, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.207016] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: bb418d62-d47c-42c5-908e-26f1eb034e46] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 777.314465] env[61857]: DEBUG nova.compute.manager [req-83071475-bf80-46b9-9881-eaac6a60d7df req-966500ed-5cbd-4895-810a-82aebad69f6f service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Received event network-changed-e9861f7d-8566-414c-a73c-1e3caf722cd2 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 777.314465] env[61857]: DEBUG nova.compute.manager [req-83071475-bf80-46b9-9881-eaac6a60d7df req-966500ed-5cbd-4895-810a-82aebad69f6f service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Refreshing instance network info cache due to event network-changed-e9861f7d-8566-414c-a73c-1e3caf722cd2. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 777.314465] env[61857]: DEBUG oslo_concurrency.lockutils [req-83071475-bf80-46b9-9881-eaac6a60d7df req-966500ed-5cbd-4895-810a-82aebad69f6f service nova] Acquiring lock "refresh_cache-c4883f84-e62b-4b59-9484-270d82dc34e0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.314465] env[61857]: DEBUG oslo_concurrency.lockutils [req-83071475-bf80-46b9-9881-eaac6a60d7df req-966500ed-5cbd-4895-810a-82aebad69f6f service nova] Acquired lock "refresh_cache-c4883f84-e62b-4b59-9484-270d82dc34e0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.314465] env[61857]: DEBUG nova.network.neutron [req-83071475-bf80-46b9-9881-eaac6a60d7df req-966500ed-5cbd-4895-810a-82aebad69f6f service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Refreshing network info cache for port e9861f7d-8566-414c-a73c-1e3caf722cd2 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 777.325014] env[61857]: DEBUG nova.network.neutron [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Successfully updated port: 45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 777.393149] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.393778] env[61857]: DEBUG nova.compute.manager [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.399034] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.227s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.401131] env[61857]: INFO nova.compute.claims [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.447748] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.474923] env[61857]: DEBUG nova.network.neutron [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updated VIF entry in instance network info cache for port 90c3f890-dd3a-4d33-92c0-a6ec8632f14b. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 777.475603] env[61857]: DEBUG nova.network.neutron [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance_info_cache with network_info: [{"id": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "address": "fa:16:3e:32:96:34", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c3f890-dd", "ovs_interfaceid": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.508967] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950850, 'name': ReconfigVM_Task, 'duration_secs': 0.300866} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.509283] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 781a2790-e317-46fe-9be8-40c9e5f1f771/781a2790-e317-46fe-9be8-40c9e5f1f771.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 777.509890] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a0eb15d-02d3-467e-aeda-6a208f889031 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.516056] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 777.516056] env[61857]: value = "task-950851" [ 777.516056] env[61857]: _type = "Task" [ 777.516056] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.525169] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950851, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.651155] env[61857]: DEBUG nova.compute.manager [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Received event network-vif-plugged-45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 777.651155] env[61857]: DEBUG oslo_concurrency.lockutils [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] Acquiring lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.651155] env[61857]: DEBUG oslo_concurrency.lockutils [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] Lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.651155] env[61857]: DEBUG oslo_concurrency.lockutils [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] Lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.651155] env[61857]: DEBUG nova.compute.manager [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] No waiting events found dispatching network-vif-plugged-45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 777.651315] env[61857]: WARNING nova.compute.manager [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Received unexpected event network-vif-plugged-45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6 for instance with vm_state building and task_state spawning. [ 777.651315] env[61857]: DEBUG nova.compute.manager [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Received event network-changed-45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 777.651315] env[61857]: DEBUG nova.compute.manager [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Refreshing instance network info cache due to event network-changed-45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 777.652802] env[61857]: DEBUG oslo_concurrency.lockutils [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] Acquiring lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.653112] env[61857]: DEBUG oslo_concurrency.lockutils [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] Acquired lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.653241] env[61857]: DEBUG nova.network.neutron [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Refreshing network info cache for port 45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 777.685490] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950849, 'name': CreateVM_Task, 'duration_secs': 0.774651} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.685664] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 777.686361] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.686516] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.686835] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 777.687399] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18900dc5-b4d5-4126-8b5e-6518bf0787d3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.692848] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 777.692848] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e11c5d-6ad1-58fe-da49-48f9ff36f1f3" [ 777.692848] env[61857]: _type = "Task" [ 777.692848] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.702017] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e11c5d-6ad1-58fe-da49-48f9ff36f1f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.710676] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 777.712728] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Cleaning up deleted instances with incomplete migration {{(pid=61857) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 777.828597] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.906032] env[61857]: DEBUG nova.compute.utils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.907222] env[61857]: DEBUG nova.compute.manager [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 777.907424] env[61857]: DEBUG nova.network.neutron [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 777.951820] env[61857]: DEBUG nova.policy [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c21b2b2013cd40a69cc7a958a821e377', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ff67efd6404381a52a4001ec774a72', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 777.978652] env[61857]: DEBUG oslo_concurrency.lockutils [req-663c010f-e5a0-4708-9f3d-376bf65c0a1d req-c0eaaf37-fd54-4d6c-bbf3-2f6d1c4f575b service nova] Releasing lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.029378] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950851, 'name': Rename_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.060407] env[61857]: DEBUG nova.network.neutron [req-83071475-bf80-46b9-9881-eaac6a60d7df req-966500ed-5cbd-4895-810a-82aebad69f6f service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Updated VIF entry in instance network info cache for port e9861f7d-8566-414c-a73c-1e3caf722cd2. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 778.060986] env[61857]: DEBUG nova.network.neutron [req-83071475-bf80-46b9-9881-eaac6a60d7df req-966500ed-5cbd-4895-810a-82aebad69f6f service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Updating instance_info_cache with network_info: [{"id": "e9861f7d-8566-414c-a73c-1e3caf722cd2", "address": "fa:16:3e:4d:95:17", "network": {"id": "3a7e17fb-072f-4bb8-87df-b69f668b4ee0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-188702091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "02a79c0554e54bc3bbd5556b890e8f93", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e1a5c1-4ae7-409b-8de7-d401684ef60d", "external-id": "nsx-vlan-transportzone-740", "segmentation_id": 740, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9861f7d-85", "ovs_interfaceid": "e9861f7d-8566-414c-a73c-1e3caf722cd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.204994] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e11c5d-6ad1-58fe-da49-48f9ff36f1f3, 'name': SearchDatastore_Task, 'duration_secs': 0.00977} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.205683] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.205946] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 778.206208] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.206384] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.206646] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 778.206941] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55c1ba2b-247a-48be-b2bd-c3d51b88bd33 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.212785] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 778.215820] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 778.216202] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 778.216863] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f5e7411-4d80-4836-b568-abda248a7c9f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.222091] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 778.222091] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5289b694-887b-2560-d791-5689302d406a" [ 778.222091] env[61857]: _type = "Task" [ 778.222091] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.230553] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5289b694-887b-2560-d791-5689302d406a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.255437] env[61857]: DEBUG nova.network.neutron [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.370198] env[61857]: DEBUG nova.network.neutron [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.412532] env[61857]: DEBUG nova.compute.manager [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 778.535321] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950851, 'name': Rename_Task, 'duration_secs': 1.011114} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.536866] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 778.536866] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-393ee378-dbe3-4122-81ee-4d475d1f55a7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.542292] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 778.542292] env[61857]: value = "task-950852" [ 778.542292] env[61857]: _type = "Task" [ 778.542292] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.547293] env[61857]: DEBUG nova.network.neutron [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Successfully created port: f1227034-4a11-4433-ae59-a08bd1100d11 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.554719] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950852, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.566228] env[61857]: DEBUG oslo_concurrency.lockutils [req-83071475-bf80-46b9-9881-eaac6a60d7df req-966500ed-5cbd-4895-810a-82aebad69f6f service nova] Releasing lock "refresh_cache-c4883f84-e62b-4b59-9484-270d82dc34e0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.733378] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5289b694-887b-2560-d791-5689302d406a, 'name': SearchDatastore_Task, 'duration_secs': 0.007716} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.737889] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23f69060-3e52-4712-9a30-0adb2914670b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.742947] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 778.742947] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]529a8963-144d-251d-555d-8ceeaf791378" [ 778.742947] env[61857]: _type = "Task" [ 778.742947] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.753630] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529a8963-144d-251d-555d-8ceeaf791378, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.873201] env[61857]: DEBUG oslo_concurrency.lockutils [req-236011ce-69f2-4281-806b-177c050a621d req-0a5e7814-5429-41c4-87cd-5d2545ac9a79 service nova] Releasing lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.874176] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.875780] env[61857]: DEBUG nova.network.neutron [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 778.940374] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9098e6bb-c1b5-4fc3-8395-71e26f817b20 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.948452] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5145a29c-b295-4a55-8c65-2b9af6908e4c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.983048] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0dab66d-839c-4015-9598-581a2847e6a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.993903] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ae2a87-3265-4da7-a82b-f0e2bb60d855 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.011410] env[61857]: DEBUG nova.compute.provider_tree [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.053020] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950852, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.254323] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529a8963-144d-251d-555d-8ceeaf791378, 'name': SearchDatastore_Task, 'duration_secs': 0.008356} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.254607] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.254872] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 71920a62-090d-4df1-937e-87df3b043e28/71920a62-090d-4df1-937e-87df3b043e28.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 779.255138] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-479e7c21-10fd-4076-b839-870512f74474 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.264026] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 779.264026] env[61857]: value = "task-950853" [ 779.264026] env[61857]: _type = "Task" [ 779.264026] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.269678] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.412540] env[61857]: DEBUG nova.network.neutron [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.430431] env[61857]: DEBUG nova.compute.manager [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.481137] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.481527] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.481722] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.481951] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.482317] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.482537] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.484127] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.484127] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.484127] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.484127] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.484127] env[61857]: DEBUG nova.virt.hardware [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.485196] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ab5283-50e0-43c2-9cbb-eaf9531e9e8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.495934] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d9411c-98cf-4980-89cf-451de93b4a3a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.517418] env[61857]: DEBUG nova.scheduler.client.report [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.554799] env[61857]: DEBUG oslo_vmware.api [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-950852, 'name': PowerOnVM_Task, 'duration_secs': 0.557751} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.557672] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 779.558076] env[61857]: INFO nova.compute.manager [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Took 8.09 seconds to spawn the instance on the hypervisor. [ 779.558352] env[61857]: DEBUG nova.compute.manager [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 779.559521] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa46f83-b133-4e90-8726-22a3d53e38a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.709554] env[61857]: DEBUG nova.network.neutron [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Updating instance_info_cache with network_info: [{"id": "45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6", "address": "fa:16:3e:05:b6:52", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45483cb0-62", "ovs_interfaceid": "45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.720953] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 779.721232] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 779.721398] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 779.773935] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950853, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438262} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.773935] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 71920a62-090d-4df1-937e-87df3b043e28/71920a62-090d-4df1-937e-87df3b043e28.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 779.774139] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 779.774381] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-77443761-73dd-41b0-857d-6e68532495de {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.781567] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 779.781567] env[61857]: value = "task-950854" [ 779.781567] env[61857]: _type = "Task" [ 779.781567] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.789326] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950854, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.025646] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.627s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.026181] env[61857]: DEBUG nova.compute.manager [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 780.033231] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.784s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.034654] env[61857]: INFO nova.compute.claims [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.081457] env[61857]: INFO nova.compute.manager [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Took 41.34 seconds to build instance. [ 780.157799] env[61857]: DEBUG nova.compute.manager [req-86046303-f033-4f7e-af33-d532d4bb952c req-f67c2f99-4fb5-484c-b962-f301a282f1c6 service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Received event network-vif-plugged-f1227034-4a11-4433-ae59-a08bd1100d11 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 780.158036] env[61857]: DEBUG oslo_concurrency.lockutils [req-86046303-f033-4f7e-af33-d532d4bb952c req-f67c2f99-4fb5-484c-b962-f301a282f1c6 service nova] Acquiring lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.158280] env[61857]: DEBUG oslo_concurrency.lockutils [req-86046303-f033-4f7e-af33-d532d4bb952c req-f67c2f99-4fb5-484c-b962-f301a282f1c6 service nova] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.158463] env[61857]: DEBUG oslo_concurrency.lockutils [req-86046303-f033-4f7e-af33-d532d4bb952c req-f67c2f99-4fb5-484c-b962-f301a282f1c6 service nova] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.158637] env[61857]: DEBUG nova.compute.manager [req-86046303-f033-4f7e-af33-d532d4bb952c req-f67c2f99-4fb5-484c-b962-f301a282f1c6 service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] No waiting events found dispatching network-vif-plugged-f1227034-4a11-4433-ae59-a08bd1100d11 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 780.158801] env[61857]: WARNING nova.compute.manager [req-86046303-f033-4f7e-af33-d532d4bb952c req-f67c2f99-4fb5-484c-b962-f301a282f1c6 service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Received unexpected event network-vif-plugged-f1227034-4a11-4433-ae59-a08bd1100d11 for instance with vm_state building and task_state spawning. [ 780.212582] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.213270] env[61857]: DEBUG nova.compute.manager [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Instance network_info: |[{"id": "45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6", "address": "fa:16:3e:05:b6:52", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45483cb0-62", "ovs_interfaceid": "45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 780.213671] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:b6:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6edb8eae-1113-49d0-84f7-9fd9f82b26fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 780.222603] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Creating folder: Project (3eca39f9d1f244a6b73b33326f4eea7c). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 780.222887] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b3639213-f22d-43bf-99b6-f82d9fb1b573 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.240753] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Created folder: Project (3eca39f9d1f244a6b73b33326f4eea7c) in parent group-v214027. [ 780.240946] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Creating folder: Instances. Parent ref: group-v214082. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 780.241231] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5e0cbf47-b87f-46ae-ba07-89b55cd9922f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.250540] env[61857]: DEBUG nova.network.neutron [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Successfully updated port: f1227034-4a11-4433-ae59-a08bd1100d11 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 780.252333] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Created folder: Instances in parent group-v214082. [ 780.252551] env[61857]: DEBUG oslo.service.loopingcall [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.254164] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 780.254331] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67158cb2-c134-459a-814d-74442d0e420c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.273702] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.273702] env[61857]: value = "task-950857" [ 780.273702] env[61857]: _type = "Task" [ 780.273702] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.281030] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950857, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.290590] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950854, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06388} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.290590] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 780.291366] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074085e8-0d83-4496-8b39-a5c3dc5df15a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.315920] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 71920a62-090d-4df1-937e-87df3b043e28/71920a62-090d-4df1-937e-87df3b043e28.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.316195] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd08690f-a9bc-4a88-ba76-15404b6f1efe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.335227] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 780.335227] env[61857]: value = "task-950858" [ 780.335227] env[61857]: _type = "Task" [ 780.335227] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.343613] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950858, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.539823] env[61857]: DEBUG nova.compute.utils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.543406] env[61857]: DEBUG nova.compute.manager [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.543580] env[61857]: DEBUG nova.network.neutron [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 780.588175] env[61857]: DEBUG oslo_concurrency.lockutils [None req-096ceb61-3e84-4938-8cc2-2a8fe0dab2b1 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.016s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.623776] env[61857]: DEBUG nova.policy [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a00d6256b5f4f78bc0fd160605123bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3eca39f9d1f244a6b73b33326f4eea7c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.755314] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "refresh_cache-36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.756775] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquired lock "refresh_cache-36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.756775] env[61857]: DEBUG nova.network.neutron [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 780.783584] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950857, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.786141] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.786274] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquired lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.786490] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Forcefully refreshing network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 780.855704] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950858, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.015951] env[61857]: DEBUG nova.network.neutron [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Successfully created port: b2d821dc-4eb3-4cb4-962e-7b235495b9be {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.044098] env[61857]: DEBUG nova.compute.manager [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.094925] env[61857]: DEBUG nova.compute.manager [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.284133] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950857, 'name': CreateVM_Task, 'duration_secs': 0.791905} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.286629] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 781.287492] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.287661] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.287974] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 781.292308] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4351ac29-a240-497b-84cc-1de63ceb6fe2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.297787] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 781.297787] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ae61dd-854f-9c43-da0c-d9aaab49b985" [ 781.297787] env[61857]: _type = "Task" [ 781.297787] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.308688] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ae61dd-854f-9c43-da0c-d9aaab49b985, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.354359] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950858, 'name': ReconfigVM_Task, 'duration_secs': 0.760063} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.354359] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 71920a62-090d-4df1-937e-87df3b043e28/71920a62-090d-4df1-937e-87df3b043e28.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.357565] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b592d3a0-4bdf-4973-b207-8ada91a42e57 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.366460] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 781.366460] env[61857]: value = "task-950859" [ 781.366460] env[61857]: _type = "Task" [ 781.366460] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.375133] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950859, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.453774] env[61857]: DEBUG nova.network.neutron [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.607740] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515483db-98e0-453d-9a39-566767e6d545 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.618984] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25993170-fd51-4b4b-b9d2-61390aff2ba8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.652157] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.655164] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76bb0973-998c-4747-a0b2-e9f1f3f7a15a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.666313] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05eeeec6-50d5-4486-bfb1-eb0400be9861 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.682540] env[61857]: DEBUG nova.compute.provider_tree [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.695237] env[61857]: DEBUG nova.network.neutron [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Updating instance_info_cache with network_info: [{"id": "f1227034-4a11-4433-ae59-a08bd1100d11", "address": "fa:16:3e:92:b7:d5", "network": {"id": "aa2712a7-cfd9-4139-8b6d-508cc240cee4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-131420680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ff67efd6404381a52a4001ec774a72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfbfc55d-8126-40dd-998e-8600ea92f97c", "external-id": "nsx-vlan-transportzone-650", "segmentation_id": 650, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1227034-4a", "ovs_interfaceid": "f1227034-4a11-4433-ae59-a08bd1100d11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.810313] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ae61dd-854f-9c43-da0c-d9aaab49b985, 'name': SearchDatastore_Task, 'duration_secs': 0.011711} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.810572] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.810798] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.811084] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.811236] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.811430] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.811709] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5800d74d-a2cf-41c9-a0ed-9ef98c086cc1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.820381] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.820573] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 781.821399] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e00abeb7-fe81-4293-aa0d-447704eb9b7f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.827891] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 781.827891] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52861b46-e947-de52-c32e-3a24a0179e39" [ 781.827891] env[61857]: _type = "Task" [ 781.827891] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.836307] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52861b46-e947-de52-c32e-3a24a0179e39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.879757] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950859, 'name': Rename_Task, 'duration_secs': 0.324072} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.880437] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 781.880725] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b58a5970-3ce6-4ca3-885c-297a89589cda {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.889056] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 781.889056] env[61857]: value = "task-950860" [ 781.889056] env[61857]: _type = "Task" [ 781.889056] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.898402] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950860, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.057028] env[61857]: DEBUG nova.compute.manager [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.122062] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.122326] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.122509] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.122757] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.122918] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.123093] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.123304] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.125159] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.125159] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.125159] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.125159] env[61857]: DEBUG nova.virt.hardware [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.126307] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20007c45-ebd0-48af-8fa2-21426ccba760 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.137697] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad2b325-e9ac-46ad-9f4e-9b15df8b1113 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.187442] env[61857]: DEBUG nova.scheduler.client.report [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.192795] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Updating instance_info_cache with network_info: [{"id": "e656ed86-dc81-4383-8dcd-254c4a462021", "address": "fa:16:3e:14:73:a6", "network": {"id": "32c36dba-b78e-4e41-ae8c-40f4b673555f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1785406565-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e0848df81fc84d25a4bb79d6f0b4b7cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape656ed86-dc", "ovs_interfaceid": "e656ed86-dc81-4383-8dcd-254c4a462021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.197835] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Releasing lock "refresh_cache-36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.197927] env[61857]: DEBUG nova.compute.manager [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Instance network_info: |[{"id": "f1227034-4a11-4433-ae59-a08bd1100d11", "address": "fa:16:3e:92:b7:d5", "network": {"id": "aa2712a7-cfd9-4139-8b6d-508cc240cee4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-131420680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ff67efd6404381a52a4001ec774a72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfbfc55d-8126-40dd-998e-8600ea92f97c", "external-id": "nsx-vlan-transportzone-650", "segmentation_id": 650, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1227034-4a", "ovs_interfaceid": "f1227034-4a11-4433-ae59-a08bd1100d11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 782.198326] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:b7:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bfbfc55d-8126-40dd-998e-8600ea92f97c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1227034-4a11-4433-ae59-a08bd1100d11', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.206330] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Creating folder: Project (38ff67efd6404381a52a4001ec774a72). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 782.207241] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-32c26d54-a0d8-4ad5-9183-e5aa43d292aa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.223054] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Created folder: Project (38ff67efd6404381a52a4001ec774a72) in parent group-v214027. [ 782.223054] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Creating folder: Instances. Parent ref: group-v214085. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 782.223054] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b0c9c54-dd0f-4edf-b65e-81ef475e5016 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.230938] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Created folder: Instances in parent group-v214085. [ 782.230938] env[61857]: DEBUG oslo.service.loopingcall [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.230938] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 782.230938] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0d8ad8c-98c0-43cd-be1b-de3a66d4150f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.250807] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.250807] env[61857]: value = "task-950863" [ 782.250807] env[61857]: _type = "Task" [ 782.250807] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.256027] env[61857]: DEBUG nova.compute.manager [req-0b7ed374-7847-4e7b-86c7-9fdff965e100 req-58491f93-e8e9-4075-9e88-3b2d9dda8861 service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Received event network-changed-f1227034-4a11-4433-ae59-a08bd1100d11 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 782.256236] env[61857]: DEBUG nova.compute.manager [req-0b7ed374-7847-4e7b-86c7-9fdff965e100 req-58491f93-e8e9-4075-9e88-3b2d9dda8861 service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Refreshing instance network info cache due to event network-changed-f1227034-4a11-4433-ae59-a08bd1100d11. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 782.256451] env[61857]: DEBUG oslo_concurrency.lockutils [req-0b7ed374-7847-4e7b-86c7-9fdff965e100 req-58491f93-e8e9-4075-9e88-3b2d9dda8861 service nova] Acquiring lock "refresh_cache-36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.256592] env[61857]: DEBUG oslo_concurrency.lockutils [req-0b7ed374-7847-4e7b-86c7-9fdff965e100 req-58491f93-e8e9-4075-9e88-3b2d9dda8861 service nova] Acquired lock "refresh_cache-36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.256752] env[61857]: DEBUG nova.network.neutron [req-0b7ed374-7847-4e7b-86c7-9fdff965e100 req-58491f93-e8e9-4075-9e88-3b2d9dda8861 service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Refreshing network info cache for port f1227034-4a11-4433-ae59-a08bd1100d11 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.264597] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950863, 'name': CreateVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.340191] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52861b46-e947-de52-c32e-3a24a0179e39, 'name': SearchDatastore_Task, 'duration_secs': 0.008882} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.341328] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-871eea16-4c2b-4ece-9184-a904bc197894 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.346900] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 782.346900] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b08a14-f796-3b02-749a-743e21833478" [ 782.346900] env[61857]: _type = "Task" [ 782.346900] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.355997] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b08a14-f796-3b02-749a-743e21833478, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.403217] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950860, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.508560] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "ddeed65b-9003-443f-8b2b-0756fbe7d234" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.508821] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "ddeed65b-9003-443f-8b2b-0756fbe7d234" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.692685] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.660s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.693402] env[61857]: DEBUG nova.compute.manager [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 782.696838] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.232s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.698325] env[61857]: INFO nova.compute.claims [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.701029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Releasing lock "refresh_cache-11adaee5-8e56-4679-8373-5d5690a44ca9" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.701270] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Updated the network info_cache for instance {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 782.701689] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.701911] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.702047] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.702206] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.702376] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.702515] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.702657] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 782.702809] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.769753] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950863, 'name': CreateVM_Task, 'duration_secs': 0.460336} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.770949] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 782.770949] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.770949] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.771437] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.771686] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e73e075-e833-4c8f-ad35-752e575feaad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.778400] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 782.778400] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5275089a-d7ed-1801-b3e1-85e02774773b" [ 782.778400] env[61857]: _type = "Task" [ 782.778400] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.790263] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5275089a-d7ed-1801-b3e1-85e02774773b, 'name': SearchDatastore_Task, 'duration_secs': 0.009349} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.790577] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.790812] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 782.791405] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.859098] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b08a14-f796-3b02-749a-743e21833478, 'name': SearchDatastore_Task, 'duration_secs': 0.009333} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.859381] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.859644] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 28bf59ca-4ffe-4005-9a88-da0660ebb48a/28bf59ca-4ffe-4005-9a88-da0660ebb48a.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 782.859928] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.860176] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 782.860435] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-00d042d3-0202-48ef-8fdd-8bcd62d09f82 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.866552] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b82db24-fbb5-4dc1-ab03-4c10ecb673a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.872025] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 782.872025] env[61857]: value = "task-950864" [ 782.872025] env[61857]: _type = "Task" [ 782.872025] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.873873] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 782.874193] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 782.876116] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-947908cb-d660-4ffb-b5a5-8b5c38a30868 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.882715] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950864, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.888775] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 782.888775] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5234dbf2-d1ce-df16-fad2-82a80d6d6142" [ 782.888775] env[61857]: _type = "Task" [ 782.888775] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.900306] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5234dbf2-d1ce-df16-fad2-82a80d6d6142, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.903158] env[61857]: DEBUG oslo_vmware.api [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950860, 'name': PowerOnVM_Task, 'duration_secs': 0.67136} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.903227] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 782.903395] env[61857]: INFO nova.compute.manager [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Took 8.11 seconds to spawn the instance on the hypervisor. [ 782.903566] env[61857]: DEBUG nova.compute.manager [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.904372] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c784e4c-e7db-4367-9505-ca72fc1c5c61 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.977909] env[61857]: DEBUG nova.network.neutron [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Successfully updated port: b2d821dc-4eb3-4cb4-962e-7b235495b9be {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 783.067527] env[61857]: DEBUG nova.network.neutron [req-0b7ed374-7847-4e7b-86c7-9fdff965e100 req-58491f93-e8e9-4075-9e88-3b2d9dda8861 service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Updated VIF entry in instance network info cache for port f1227034-4a11-4433-ae59-a08bd1100d11. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.068162] env[61857]: DEBUG nova.network.neutron [req-0b7ed374-7847-4e7b-86c7-9fdff965e100 req-58491f93-e8e9-4075-9e88-3b2d9dda8861 service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Updating instance_info_cache with network_info: [{"id": "f1227034-4a11-4433-ae59-a08bd1100d11", "address": "fa:16:3e:92:b7:d5", "network": {"id": "aa2712a7-cfd9-4139-8b6d-508cc240cee4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-131420680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ff67efd6404381a52a4001ec774a72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfbfc55d-8126-40dd-998e-8600ea92f97c", "external-id": "nsx-vlan-transportzone-650", "segmentation_id": 650, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1227034-4a", "ovs_interfaceid": "f1227034-4a11-4433-ae59-a08bd1100d11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.205668] env[61857]: DEBUG nova.compute.utils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.207659] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.208011] env[61857]: DEBUG nova.compute.manager [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.208192] env[61857]: DEBUG nova.network.neutron [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 783.258973] env[61857]: DEBUG nova.policy [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac8387df3290404283263440672de653', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3bbcf7c2937040e1906e2273a07b671b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 783.269675] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquiring lock "11adaee5-8e56-4679-8373-5d5690a44ca9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.269974] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lock "11adaee5-8e56-4679-8373-5d5690a44ca9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.270190] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquiring lock "11adaee5-8e56-4679-8373-5d5690a44ca9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.270404] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lock "11adaee5-8e56-4679-8373-5d5690a44ca9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.270558] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lock "11adaee5-8e56-4679-8373-5d5690a44ca9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.272997] env[61857]: INFO nova.compute.manager [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Terminating instance [ 783.275180] env[61857]: DEBUG nova.compute.manager [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 783.275379] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 783.276272] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578152fb-b45f-4de5-8a78-780209a7aff7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.285988] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 783.286306] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a37d95bf-fd88-44ff-be37-f4f0a95bfa39 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.295382] env[61857]: DEBUG oslo_vmware.api [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for the task: (returnval){ [ 783.295382] env[61857]: value = "task-950865" [ 783.295382] env[61857]: _type = "Task" [ 783.295382] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.307567] env[61857]: DEBUG oslo_vmware.api [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.378890] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950864, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.400353] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5234dbf2-d1ce-df16-fad2-82a80d6d6142, 'name': SearchDatastore_Task, 'duration_secs': 0.009089} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.401440] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41a12e95-8143-404d-b991-6b43b89a576a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.408288] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 783.408288] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52a5afec-e2ec-6a28-8a62-968e2e13185f" [ 783.408288] env[61857]: _type = "Task" [ 783.408288] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.426130] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a5afec-e2ec-6a28-8a62-968e2e13185f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.426130] env[61857]: INFO nova.compute.manager [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Took 38.73 seconds to build instance. [ 783.483729] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "refresh_cache-e91e99b9-1cd3-4345-af09-f14af4df1214" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.483729] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "refresh_cache-e91e99b9-1cd3-4345-af09-f14af4df1214" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.483729] env[61857]: DEBUG nova.network.neutron [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 783.573064] env[61857]: DEBUG oslo_concurrency.lockutils [req-0b7ed374-7847-4e7b-86c7-9fdff965e100 req-58491f93-e8e9-4075-9e88-3b2d9dda8861 service nova] Releasing lock "refresh_cache-36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.715259] env[61857]: DEBUG nova.network.neutron [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Successfully created port: 9d8efcbb-1103-49f6-a52f-219e1682d275 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 783.723581] env[61857]: DEBUG nova.compute.manager [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.810980] env[61857]: DEBUG oslo_vmware.api [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950865, 'name': PowerOffVM_Task, 'duration_secs': 0.507971} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.813537] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 783.813762] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 783.814343] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51e0e287-0852-42a5-943a-0615834f2b9f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.882749] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950864, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528773} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.885471] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 28bf59ca-4ffe-4005-9a88-da0660ebb48a/28bf59ca-4ffe-4005-9a88-da0660ebb48a.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 783.885700] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.887208] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8cf1e3ff-23d1-4461-9753-e416e68912eb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.896046] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 783.896046] env[61857]: value = "task-950867" [ 783.896046] env[61857]: _type = "Task" [ 783.896046] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.910471] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950867, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.921414] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a5afec-e2ec-6a28-8a62-968e2e13185f, 'name': SearchDatastore_Task, 'duration_secs': 0.023527} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.924203] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.924566] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5/36ca32ed-1ba2-40d0-93c4-086a26a3f9e5.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 783.925037] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec0a34c1-3fc0-4bc5-bb7f-4a6dd073a71c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.927501] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a1362f92-504a-4d61-acab-8680a9d8cb5c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.716s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.933702] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 783.933702] env[61857]: value = "task-950868" [ 783.933702] env[61857]: _type = "Task" [ 783.933702] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.946666] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950868, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.967351] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 783.971099] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 783.971099] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Deleting the datastore file [datastore2] 11adaee5-8e56-4679-8373-5d5690a44ca9 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 783.971099] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d51f02d-7ab0-4e36-8e46-86fd5bddcf21 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.974736] env[61857]: DEBUG oslo_vmware.api [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for the task: (returnval){ [ 783.974736] env[61857]: value = "task-950869" [ 783.974736] env[61857]: _type = "Task" [ 783.974736] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.990103] env[61857]: DEBUG oslo_vmware.api [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950869, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.055472] env[61857]: DEBUG nova.network.neutron [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.315692] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3f558b-4329-4114-b3d2-6e837314f5e8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.328718] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136cd5ad-6ecf-4a84-b3a0-e3975cdd2cd1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.372833] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15395d90-c4c7-4b69-b5c7-e916f72f39c0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.382937] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce22636b-5cfe-4fe3-8e09-008af81a5a13 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.402277] env[61857]: DEBUG nova.compute.provider_tree [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 784.414688] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950867, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.254965} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.415073] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.416677] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64518e5-d8c8-4d5e-a2e2-388b0ca0dcc6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.434747] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.444939] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 28bf59ca-4ffe-4005-9a88-da0660ebb48a/28bf59ca-4ffe-4005-9a88-da0660ebb48a.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.447152] env[61857]: DEBUG nova.network.neutron [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Updating instance_info_cache with network_info: [{"id": "b2d821dc-4eb3-4cb4-962e-7b235495b9be", "address": "fa:16:3e:60:87:33", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2d821dc-4e", "ovs_interfaceid": "b2d821dc-4eb3-4cb4-962e-7b235495b9be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.448293] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9d82175-7994-47ed-8bb1-3c353de0bc43 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.474365] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950868, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.476858] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 784.476858] env[61857]: value = "task-950870" [ 784.476858] env[61857]: _type = "Task" [ 784.476858] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.489032] env[61857]: DEBUG oslo_vmware.api [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Task: {'id': task-950869, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264929} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.492334] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 784.492546] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 784.492714] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 784.492889] env[61857]: INFO nova.compute.manager [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Took 1.22 seconds to destroy the instance on the hypervisor. [ 784.493172] env[61857]: DEBUG oslo.service.loopingcall [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.493383] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950870, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.494220] env[61857]: DEBUG nova.compute.manager [-] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.494316] env[61857]: DEBUG nova.network.neutron [-] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 784.554088] env[61857]: DEBUG nova.compute.manager [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Received event network-vif-plugged-b2d821dc-4eb3-4cb4-962e-7b235495b9be {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.554595] env[61857]: DEBUG oslo_concurrency.lockutils [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] Acquiring lock "e91e99b9-1cd3-4345-af09-f14af4df1214-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.554595] env[61857]: DEBUG oslo_concurrency.lockutils [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] Lock "e91e99b9-1cd3-4345-af09-f14af4df1214-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.554733] env[61857]: DEBUG oslo_concurrency.lockutils [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] Lock "e91e99b9-1cd3-4345-af09-f14af4df1214-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.554830] env[61857]: DEBUG nova.compute.manager [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] No waiting events found dispatching network-vif-plugged-b2d821dc-4eb3-4cb4-962e-7b235495b9be {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 784.555015] env[61857]: WARNING nova.compute.manager [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Received unexpected event network-vif-plugged-b2d821dc-4eb3-4cb4-962e-7b235495b9be for instance with vm_state building and task_state spawning. [ 784.555181] env[61857]: DEBUG nova.compute.manager [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Received event network-changed-b2d821dc-4eb3-4cb4-962e-7b235495b9be {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.555339] env[61857]: DEBUG nova.compute.manager [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Refreshing instance network info cache due to event network-changed-b2d821dc-4eb3-4cb4-962e-7b235495b9be. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 784.555509] env[61857]: DEBUG oslo_concurrency.lockutils [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] Acquiring lock "refresh_cache-e91e99b9-1cd3-4345-af09-f14af4df1214" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.739238] env[61857]: DEBUG nova.compute.manager [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 784.796561] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.796946] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.797223] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.797527] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.797770] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.798139] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.798365] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.798627] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.798897] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.799177] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.799461] env[61857]: DEBUG nova.virt.hardware [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.800764] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8725e34b-ef2c-4c33-ac3f-ff0b358555d8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.812197] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b524b613-0e96-4cc7-9858-c366d2314365 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.927350] env[61857]: ERROR nova.scheduler.client.report [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [req-de88ccc6-c5f7-45d7-86d4-7872799ed787] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2d5860d0-8574-4e55-9ced-20e66f0314c2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-de88ccc6-c5f7-45d7-86d4-7872799ed787"}]} [ 784.943661] env[61857]: DEBUG nova.scheduler.client.report [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Refreshing inventories for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 784.960621] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950868, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.636416} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.960910] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5/36ca32ed-1ba2-40d0-93c4-086a26a3f9e5.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 784.961142] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.961404] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0fb65dbc-fabf-40c2-8c23-60160bb85d84 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.964363] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.964796] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "refresh_cache-e91e99b9-1cd3-4345-af09-f14af4df1214" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.965118] env[61857]: DEBUG nova.compute.manager [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Instance network_info: |[{"id": "b2d821dc-4eb3-4cb4-962e-7b235495b9be", "address": "fa:16:3e:60:87:33", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2d821dc-4e", "ovs_interfaceid": "b2d821dc-4eb3-4cb4-962e-7b235495b9be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 784.966094] env[61857]: DEBUG nova.scheduler.client.report [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Updating ProviderTree inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 784.966301] env[61857]: DEBUG nova.compute.provider_tree [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 784.968289] env[61857]: DEBUG oslo_concurrency.lockutils [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] Acquired lock "refresh_cache-e91e99b9-1cd3-4345-af09-f14af4df1214" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.968502] env[61857]: DEBUG nova.network.neutron [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Refreshing network info cache for port b2d821dc-4eb3-4cb4-962e-7b235495b9be {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.969756] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:87:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6edb8eae-1113-49d0-84f7-9fd9f82b26fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2d821dc-4eb3-4cb4-962e-7b235495b9be', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 784.977924] env[61857]: DEBUG oslo.service.loopingcall [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.980506] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 784.980784] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 784.980784] env[61857]: value = "task-950871" [ 784.980784] env[61857]: _type = "Task" [ 784.980784] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.981421] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b8c17ef-68b2-4742-ad3b-5adf182ce268 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.999455] env[61857]: DEBUG nova.scheduler.client.report [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Refreshing aggregate associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, aggregates: None {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 785.013908] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950870, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.018486] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.018486] env[61857]: value = "task-950872" [ 785.018486] env[61857]: _type = "Task" [ 785.018486] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.018486] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950871, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.029269] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950872, 'name': CreateVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.043880] env[61857]: DEBUG nova.scheduler.client.report [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Refreshing trait associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 785.412820] env[61857]: DEBUG nova.compute.manager [req-0a799ce8-cf89-4437-99a8-13730b5ecc4c req-5a03cf6f-8917-4df8-a3c6-50969b89cbe8 service nova] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Received event network-vif-plugged-9d8efcbb-1103-49f6-a52f-219e1682d275 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 785.413092] env[61857]: DEBUG oslo_concurrency.lockutils [req-0a799ce8-cf89-4437-99a8-13730b5ecc4c req-5a03cf6f-8917-4df8-a3c6-50969b89cbe8 service nova] Acquiring lock "30d905ed-831d-44ae-807c-062de9a7e9fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.413441] env[61857]: DEBUG oslo_concurrency.lockutils [req-0a799ce8-cf89-4437-99a8-13730b5ecc4c req-5a03cf6f-8917-4df8-a3c6-50969b89cbe8 service nova] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.413478] env[61857]: DEBUG oslo_concurrency.lockutils [req-0a799ce8-cf89-4437-99a8-13730b5ecc4c req-5a03cf6f-8917-4df8-a3c6-50969b89cbe8 service nova] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.413643] env[61857]: DEBUG nova.compute.manager [req-0a799ce8-cf89-4437-99a8-13730b5ecc4c req-5a03cf6f-8917-4df8-a3c6-50969b89cbe8 service nova] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] No waiting events found dispatching network-vif-plugged-9d8efcbb-1103-49f6-a52f-219e1682d275 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 785.413806] env[61857]: WARNING nova.compute.manager [req-0a799ce8-cf89-4437-99a8-13730b5ecc4c req-5a03cf6f-8917-4df8-a3c6-50969b89cbe8 service nova] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Received unexpected event network-vif-plugged-9d8efcbb-1103-49f6-a52f-219e1682d275 for instance with vm_state building and task_state spawning. [ 785.495119] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950870, 'name': ReconfigVM_Task, 'duration_secs': 0.724879} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.495425] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 28bf59ca-4ffe-4005-9a88-da0660ebb48a/28bf59ca-4ffe-4005-9a88-da0660ebb48a.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 785.496048] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3419f79-be48-414c-b10e-d9524a1165bb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.505395] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 785.505395] env[61857]: value = "task-950873" [ 785.505395] env[61857]: _type = "Task" [ 785.505395] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.510452] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950871, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137224} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.511061] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.511888] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66114436-3dba-4d05-ab38-fb5cc61e4416 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.521040] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950873, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.541287] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5/36ca32ed-1ba2-40d0-93c4-086a26a3f9e5.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.545210] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53bfb0b7-0126-496d-ac3a-62bcdd61cddd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.561470] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064d6daf-b679-4768-acb3-31df2b87015a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.563318] env[61857]: DEBUG nova.network.neutron [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Successfully updated port: 9d8efcbb-1103-49f6-a52f-219e1682d275 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 785.568429] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950872, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.574357] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14917ed4-7fb5-4110-84e9-21661698fab7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.578421] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 785.578421] env[61857]: value = "task-950874" [ 785.578421] env[61857]: _type = "Task" [ 785.578421] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.609872] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1f45a7-cc58-4bb4-9e17-7373f032d277 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.616507] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950874, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.621614] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2dce8c-3a90-4a43-91bf-80a4f6db5ab2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.637408] env[61857]: DEBUG nova.compute.provider_tree [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 785.737370] env[61857]: DEBUG nova.network.neutron [-] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.827649] env[61857]: DEBUG nova.network.neutron [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Updated VIF entry in instance network info cache for port b2d821dc-4eb3-4cb4-962e-7b235495b9be. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 785.828124] env[61857]: DEBUG nova.network.neutron [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Updating instance_info_cache with network_info: [{"id": "b2d821dc-4eb3-4cb4-962e-7b235495b9be", "address": "fa:16:3e:60:87:33", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2d821dc-4e", "ovs_interfaceid": "b2d821dc-4eb3-4cb4-962e-7b235495b9be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.015153] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950873, 'name': Rename_Task, 'duration_secs': 0.14629} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.015480] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 786.015725] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c13f3c04-1c87-41da-9aeb-194d8a8af5cc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.024462] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 786.024462] env[61857]: value = "task-950875" [ 786.024462] env[61857]: _type = "Task" [ 786.024462] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.033652] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950872, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.036593] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950875, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.070493] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.070640] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.070841] env[61857]: DEBUG nova.network.neutron [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 786.088886] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950874, 'name': ReconfigVM_Task, 'duration_secs': 0.281886} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.089193] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5/36ca32ed-1ba2-40d0-93c4-086a26a3f9e5.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.089870] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7a26717-e001-4ef9-bf89-df9e17ddb6c2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.095790] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 786.095790] env[61857]: value = "task-950876" [ 786.095790] env[61857]: _type = "Task" [ 786.095790] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.103622] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950876, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.244423] env[61857]: INFO nova.compute.manager [-] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Took 1.75 seconds to deallocate network for instance. [ 786.334760] env[61857]: DEBUG oslo_concurrency.lockutils [req-5156647f-4239-46c5-81ea-a0452bc62870 req-2d2a6010-8d02-47a3-8d35-d8486fa52774 service nova] Releasing lock "refresh_cache-e91e99b9-1cd3-4345-af09-f14af4df1214" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.351457] env[61857]: DEBUG nova.scheduler.client.report [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Updated inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with generation 81 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 786.351457] env[61857]: DEBUG nova.compute.provider_tree [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Updating resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 generation from 81 to 82 during operation: update_inventory {{(pid=61857) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 786.351457] env[61857]: DEBUG nova.compute.provider_tree [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.531390] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950872, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.536312] env[61857]: DEBUG oslo_vmware.api [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950875, 'name': PowerOnVM_Task, 'duration_secs': 0.469993} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.536560] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 786.536756] env[61857]: INFO nova.compute.manager [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Took 9.72 seconds to spawn the instance on the hypervisor. [ 786.536933] env[61857]: DEBUG nova.compute.manager [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 786.537713] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251dd05e-0b90-45bf-9d47-4a4b64f62355 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.587498] env[61857]: DEBUG nova.compute.manager [req-bde0134e-0d3a-4296-b4b1-c81fb3dee283 req-b23ce93b-1bea-4b4d-8663-849995895833 service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Received event network-changed-90c3f890-dd3a-4d33-92c0-a6ec8632f14b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 786.587785] env[61857]: DEBUG nova.compute.manager [req-bde0134e-0d3a-4296-b4b1-c81fb3dee283 req-b23ce93b-1bea-4b4d-8663-849995895833 service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Refreshing instance network info cache due to event network-changed-90c3f890-dd3a-4d33-92c0-a6ec8632f14b. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 786.587909] env[61857]: DEBUG oslo_concurrency.lockutils [req-bde0134e-0d3a-4296-b4b1-c81fb3dee283 req-b23ce93b-1bea-4b4d-8663-849995895833 service nova] Acquiring lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.588082] env[61857]: DEBUG oslo_concurrency.lockutils [req-bde0134e-0d3a-4296-b4b1-c81fb3dee283 req-b23ce93b-1bea-4b4d-8663-849995895833 service nova] Acquired lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.588268] env[61857]: DEBUG nova.network.neutron [req-bde0134e-0d3a-4296-b4b1-c81fb3dee283 req-b23ce93b-1bea-4b4d-8663-849995895833 service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Refreshing network info cache for port 90c3f890-dd3a-4d33-92c0-a6ec8632f14b {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 786.603143] env[61857]: DEBUG nova.network.neutron [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.608933] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950876, 'name': Rename_Task, 'duration_secs': 0.169921} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.609210] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 786.609445] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f37bba2e-7aaa-4f14-b42b-f59cb152076e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.616091] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 786.616091] env[61857]: value = "task-950877" [ 786.616091] env[61857]: _type = "Task" [ 786.616091] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.624716] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950877, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.753423] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.784195] env[61857]: DEBUG nova.network.neutron [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance_info_cache with network_info: [{"id": "9d8efcbb-1103-49f6-a52f-219e1682d275", "address": "fa:16:3e:54:79:1a", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d8efcbb-11", "ovs_interfaceid": "9d8efcbb-1103-49f6-a52f-219e1682d275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.857472] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.160s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.857626] env[61857]: DEBUG nova.compute.manager [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 786.860440] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.681s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.860684] env[61857]: DEBUG nova.objects.instance [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lazy-loading 'resources' on Instance uuid c4f70321-f612-416e-b209-eb974dab9d49 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 787.030947] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950872, 'name': CreateVM_Task, 'duration_secs': 1.852254} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.031147] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 787.031845] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.032016] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.032373] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.032611] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ebf30c8-80f8-4cd0-b5b4-821de8c4e179 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.037130] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 787.037130] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52db122d-9c28-4397-5269-e30ae63d3a35" [ 787.037130] env[61857]: _type = "Task" [ 787.037130] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.044972] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52db122d-9c28-4397-5269-e30ae63d3a35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.052689] env[61857]: INFO nova.compute.manager [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Took 39.81 seconds to build instance. [ 787.129025] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950877, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.287920] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.287920] env[61857]: DEBUG nova.compute.manager [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Instance network_info: |[{"id": "9d8efcbb-1103-49f6-a52f-219e1682d275", "address": "fa:16:3e:54:79:1a", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d8efcbb-11", "ovs_interfaceid": "9d8efcbb-1103-49f6-a52f-219e1682d275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 787.288322] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:79:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3a80436-f7a9-431a-acec-aca3d76e3f9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d8efcbb-1103-49f6-a52f-219e1682d275', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.297774] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Creating folder: Project (3bbcf7c2937040e1906e2273a07b671b). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 787.300446] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f7e7a37-7d1a-4bee-889d-8e6ed8a14ff3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.312798] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Created folder: Project (3bbcf7c2937040e1906e2273a07b671b) in parent group-v214027. [ 787.313425] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Creating folder: Instances. Parent ref: group-v214089. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 787.313425] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-40c98937-b88b-4079-bf17-732f8963f352 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.322510] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Created folder: Instances in parent group-v214089. [ 787.322801] env[61857]: DEBUG oslo.service.loopingcall [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.323015] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 787.323227] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-605768ad-99f5-452a-85f8-023bc52bfae0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.345568] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.345568] env[61857]: value = "task-950880" [ 787.345568] env[61857]: _type = "Task" [ 787.345568] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.361404] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950880, 'name': CreateVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.363768] env[61857]: DEBUG nova.compute.utils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.367709] env[61857]: DEBUG nova.compute.manager [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 787.367882] env[61857]: DEBUG nova.network.neutron [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 787.414327] env[61857]: DEBUG nova.network.neutron [req-bde0134e-0d3a-4296-b4b1-c81fb3dee283 req-b23ce93b-1bea-4b4d-8663-849995895833 service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updated VIF entry in instance network info cache for port 90c3f890-dd3a-4d33-92c0-a6ec8632f14b. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 787.414640] env[61857]: DEBUG nova.network.neutron [req-bde0134e-0d3a-4296-b4b1-c81fb3dee283 req-b23ce93b-1bea-4b4d-8663-849995895833 service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance_info_cache with network_info: [{"id": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "address": "fa:16:3e:32:96:34", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c3f890-dd", "ovs_interfaceid": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.452844] env[61857]: DEBUG nova.policy [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd71828dc4ff54dc7939df4e759285943', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b51cf69f1c947c89a3969daba09361b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.485968] env[61857]: DEBUG nova.compute.manager [req-f7f42030-fd71-41e9-9e89-f2ada8113acd req-63bfaa4c-15dd-48a5-b413-642841edc66f service nova] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Received event network-changed-9d8efcbb-1103-49f6-a52f-219e1682d275 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.486192] env[61857]: DEBUG nova.compute.manager [req-f7f42030-fd71-41e9-9e89-f2ada8113acd req-63bfaa4c-15dd-48a5-b413-642841edc66f service nova] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Refreshing instance network info cache due to event network-changed-9d8efcbb-1103-49f6-a52f-219e1682d275. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 787.486412] env[61857]: DEBUG oslo_concurrency.lockutils [req-f7f42030-fd71-41e9-9e89-f2ada8113acd req-63bfaa4c-15dd-48a5-b413-642841edc66f service nova] Acquiring lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.486558] env[61857]: DEBUG oslo_concurrency.lockutils [req-f7f42030-fd71-41e9-9e89-f2ada8113acd req-63bfaa4c-15dd-48a5-b413-642841edc66f service nova] Acquired lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.486719] env[61857]: DEBUG nova.network.neutron [req-f7f42030-fd71-41e9-9e89-f2ada8113acd req-63bfaa4c-15dd-48a5-b413-642841edc66f service nova] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Refreshing network info cache for port 9d8efcbb-1103-49f6-a52f-219e1682d275 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 787.548496] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52db122d-9c28-4397-5269-e30ae63d3a35, 'name': SearchDatastore_Task, 'duration_secs': 0.01142} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.548834] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.549050] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.549822] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.549822] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.549822] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.551144] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e000017-39d6-476d-a2ce-757cbff9886e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.554737] env[61857]: DEBUG oslo_concurrency.lockutils [None req-02446869-dd55-43ab-83f0-d8e80f9ce104 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.945s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.560353] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.560533] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 787.561437] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d29947b6-b2b5-435d-ba54-df95ae2f52a4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.566738] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 787.566738] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52093507-33b2-a7d2-6202-de7ef8d7c92b" [ 787.566738] env[61857]: _type = "Task" [ 787.566738] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.580386] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52093507-33b2-a7d2-6202-de7ef8d7c92b, 'name': SearchDatastore_Task, 'duration_secs': 0.008843} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.580689] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a5e302b-d0aa-4bb5-b663-457935733efa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.587723] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 787.587723] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52eb362c-e393-1452-7795-7ef7a92ca3c5" [ 787.587723] env[61857]: _type = "Task" [ 787.587723] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.596345] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52eb362c-e393-1452-7795-7ef7a92ca3c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.633821] env[61857]: DEBUG oslo_vmware.api [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-950877, 'name': PowerOnVM_Task, 'duration_secs': 0.930261} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.633821] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 787.633821] env[61857]: INFO nova.compute.manager [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Took 8.20 seconds to spawn the instance on the hypervisor. [ 787.633821] env[61857]: DEBUG nova.compute.manager [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 787.633821] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71cbb40-1ccd-4699-9b93-0fa0d70cdf03 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.854475] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950880, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.870121] env[61857]: DEBUG nova.compute.manager [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.918058] env[61857]: DEBUG oslo_concurrency.lockutils [req-bde0134e-0d3a-4296-b4b1-c81fb3dee283 req-b23ce93b-1bea-4b4d-8663-849995895833 service nova] Releasing lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.918058] env[61857]: DEBUG nova.compute.manager [req-bde0134e-0d3a-4296-b4b1-c81fb3dee283 req-b23ce93b-1bea-4b4d-8663-849995895833 service nova] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Received event network-vif-deleted-e656ed86-dc81-4383-8dcd-254c4a462021 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.939555] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f23f6c5-53e3-43f8-950c-56c800033130 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.943579] env[61857]: DEBUG nova.network.neutron [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Successfully created port: 25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 787.952276] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549523be-f0bf-4eb8-942f-ab5da680fdff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.988182] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c9df28-b3d4-4915-bfc2-efc27390b369 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.998913] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76e039d-df64-4961-8563-3f31ca881af1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.013382] env[61857]: DEBUG nova.compute.provider_tree [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.059616] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.098647] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52eb362c-e393-1452-7795-7ef7a92ca3c5, 'name': SearchDatastore_Task, 'duration_secs': 0.009009} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.098925] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.099196] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] e91e99b9-1cd3-4345-af09-f14af4df1214/e91e99b9-1cd3-4345-af09-f14af4df1214.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 788.099444] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0bf2dcb-c53d-4d7a-bf42-7aeb208f4a71 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.105718] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 788.105718] env[61857]: value = "task-950881" [ 788.105718] env[61857]: _type = "Task" [ 788.105718] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.113270] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950881, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.152856] env[61857]: INFO nova.compute.manager [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Took 37.24 seconds to build instance. [ 788.351376] env[61857]: DEBUG nova.network.neutron [req-f7f42030-fd71-41e9-9e89-f2ada8113acd req-63bfaa4c-15dd-48a5-b413-642841edc66f service nova] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updated VIF entry in instance network info cache for port 9d8efcbb-1103-49f6-a52f-219e1682d275. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.351808] env[61857]: DEBUG nova.network.neutron [req-f7f42030-fd71-41e9-9e89-f2ada8113acd req-63bfaa4c-15dd-48a5-b413-642841edc66f service nova] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance_info_cache with network_info: [{"id": "9d8efcbb-1103-49f6-a52f-219e1682d275", "address": "fa:16:3e:54:79:1a", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d8efcbb-11", "ovs_interfaceid": "9d8efcbb-1103-49f6-a52f-219e1682d275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.362658] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950880, 'name': CreateVM_Task, 'duration_secs': 0.553} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.362658] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 788.362658] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.362658] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.362658] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 788.362914] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21741d4b-9dcc-4805-b452-000361ae0125 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.368925] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 788.368925] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52119e51-1897-201a-d49c-11264f767406" [ 788.368925] env[61857]: _type = "Task" [ 788.368925] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.382995] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52119e51-1897-201a-d49c-11264f767406, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.517281] env[61857]: DEBUG nova.scheduler.client.report [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.599114] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.617806] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950881, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480553} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.618098] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] e91e99b9-1cd3-4345-af09-f14af4df1214/e91e99b9-1cd3-4345-af09-f14af4df1214.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 788.618322] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.618566] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e3fafb9-dad4-4fcc-b2fb-d74e497f4f39 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.624496] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 788.624496] env[61857]: value = "task-950882" [ 788.624496] env[61857]: _type = "Task" [ 788.624496] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.632268] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950882, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.654955] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da9c18c8-30b6-4531-bb38-ab4a27310682 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.827s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.863305] env[61857]: DEBUG oslo_concurrency.lockutils [req-f7f42030-fd71-41e9-9e89-f2ada8113acd req-63bfaa4c-15dd-48a5-b413-642841edc66f service nova] Releasing lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.881802] env[61857]: DEBUG nova.compute.manager [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.894600] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52119e51-1897-201a-d49c-11264f767406, 'name': SearchDatastore_Task, 'duration_secs': 0.048294} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.894973] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.895238] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.895479] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.895629] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.895808] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.896170] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb1b2e13-cf1e-4590-a040-6328a40c0370 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.905909] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.906175] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 788.906978] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2b08a41-475d-4f7c-a226-28c7524cbe0a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.913468] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 788.913468] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ef5abd-a040-55b7-1876-a0bab1174530" [ 788.913468] env[61857]: _type = "Task" [ 788.913468] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.921557] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ef5abd-a040-55b7-1876-a0bab1174530, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.943870] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.944154] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.944316] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.944500] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.944857] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.944857] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.945008] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.945269] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.945453] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.945621] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.945824] env[61857]: DEBUG nova.virt.hardware [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.947318] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f285c802-ac3d-4ff5-8819-708edc651ec5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.955721] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c528872-db82-44e5-a068-5c25d9e28386 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.025866] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.165s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.028458] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.988s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.030087] env[61857]: INFO nova.compute.claims [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.071110] env[61857]: INFO nova.scheduler.client.report [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Deleted allocations for instance c4f70321-f612-416e-b209-eb974dab9d49 [ 789.134820] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950882, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063227} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.134921] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.136667] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04e3513-3f85-4996-9b0f-6067b6cf2b8a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.166951] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] e91e99b9-1cd3-4345-af09-f14af4df1214/e91e99b9-1cd3-4345-af09-f14af4df1214.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.166951] env[61857]: DEBUG nova.compute.manager [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 789.166951] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1813206a-7a0c-4137-9b7c-482e7b735565 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.188564] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 789.188564] env[61857]: value = "task-950883" [ 789.188564] env[61857]: _type = "Task" [ 789.188564] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.200228] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950883, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.428673] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ef5abd-a040-55b7-1876-a0bab1174530, 'name': SearchDatastore_Task, 'duration_secs': 0.024719} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.429673] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b1a4d88-acde-4ac7-aca8-df1613bce064 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.435113] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 789.435113] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ce8b08-6a65-e175-cca6-3c077f61323e" [ 789.435113] env[61857]: _type = "Task" [ 789.435113] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.442779] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ce8b08-6a65-e175-cca6-3c077f61323e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.529419] env[61857]: DEBUG nova.compute.manager [req-68ba89b5-5435-4944-908d-165c35c21df3 req-8d238cb7-6e2d-4c90-afa6-3c55e33e336f service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Received event network-changed-f1227034-4a11-4433-ae59-a08bd1100d11 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 789.529419] env[61857]: DEBUG nova.compute.manager [req-68ba89b5-5435-4944-908d-165c35c21df3 req-8d238cb7-6e2d-4c90-afa6-3c55e33e336f service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Refreshing instance network info cache due to event network-changed-f1227034-4a11-4433-ae59-a08bd1100d11. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 789.529419] env[61857]: DEBUG oslo_concurrency.lockutils [req-68ba89b5-5435-4944-908d-165c35c21df3 req-8d238cb7-6e2d-4c90-afa6-3c55e33e336f service nova] Acquiring lock "refresh_cache-36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.529419] env[61857]: DEBUG oslo_concurrency.lockutils [req-68ba89b5-5435-4944-908d-165c35c21df3 req-8d238cb7-6e2d-4c90-afa6-3c55e33e336f service nova] Acquired lock "refresh_cache-36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.529419] env[61857]: DEBUG nova.network.neutron [req-68ba89b5-5435-4944-908d-165c35c21df3 req-8d238cb7-6e2d-4c90-afa6-3c55e33e336f service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Refreshing network info cache for port f1227034-4a11-4433-ae59-a08bd1100d11 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 789.584358] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c8bb6fb6-53a8-469b-b717-9fde6775c004 tempest-VolumesAdminNegativeTest-1420447228 tempest-VolumesAdminNegativeTest-1420447228-project-member] Lock "c4f70321-f612-416e-b209-eb974dab9d49" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.870s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.704080] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.728805] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.808363] env[61857]: DEBUG nova.network.neutron [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Successfully updated port: 25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 789.946926] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ce8b08-6a65-e175-cca6-3c077f61323e, 'name': SearchDatastore_Task, 'duration_secs': 0.010022} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.947250] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.947633] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 30d905ed-831d-44ae-807c-062de9a7e9fb/30d905ed-831d-44ae-807c-062de9a7e9fb.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 789.947963] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23d8edb1-9cf6-474f-ac98-afafa258e03e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.954242] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 789.954242] env[61857]: value = "task-950884" [ 789.954242] env[61857]: _type = "Task" [ 789.954242] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.962081] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.204312] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950883, 'name': ReconfigVM_Task, 'duration_secs': 0.810178} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.204782] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Reconfigured VM instance instance-00000038 to attach disk [datastore2] e91e99b9-1cd3-4345-af09-f14af4df1214/e91e99b9-1cd3-4345-af09-f14af4df1214.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.205355] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5ac5505-ba62-42ef-9a50-7bf2eccb5ac3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.213608] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 790.213608] env[61857]: value = "task-950885" [ 790.213608] env[61857]: _type = "Task" [ 790.213608] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.222977] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950885, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.315685] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.315685] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquired lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.315685] env[61857]: DEBUG nova.network.neutron [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 790.446059] env[61857]: DEBUG nova.network.neutron [req-68ba89b5-5435-4944-908d-165c35c21df3 req-8d238cb7-6e2d-4c90-afa6-3c55e33e336f service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Updated VIF entry in instance network info cache for port f1227034-4a11-4433-ae59-a08bd1100d11. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.446440] env[61857]: DEBUG nova.network.neutron [req-68ba89b5-5435-4944-908d-165c35c21df3 req-8d238cb7-6e2d-4c90-afa6-3c55e33e336f service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Updating instance_info_cache with network_info: [{"id": "f1227034-4a11-4433-ae59-a08bd1100d11", "address": "fa:16:3e:92:b7:d5", "network": {"id": "aa2712a7-cfd9-4139-8b6d-508cc240cee4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-131420680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ff67efd6404381a52a4001ec774a72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfbfc55d-8126-40dd-998e-8600ea92f97c", "external-id": "nsx-vlan-transportzone-650", "segmentation_id": 650, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1227034-4a", "ovs_interfaceid": "f1227034-4a11-4433-ae59-a08bd1100d11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.466539] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950884, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486724} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.468425] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 30d905ed-831d-44ae-807c-062de9a7e9fb/30d905ed-831d-44ae-807c-062de9a7e9fb.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 790.468682] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 790.469185] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ead949e-0508-4918-a274-fe62f9bd1bbf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.475280] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 790.475280] env[61857]: value = "task-950886" [ 790.475280] env[61857]: _type = "Task" [ 790.475280] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.489738] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950886, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.563908] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf8bcbc-7bf5-4c85-b70d-956c24b6906b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.572052] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b79e1f1-1a6f-4af9-8de6-192df0cd8c34 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.604875] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e1b9ab-7431-4abe-8b0c-559225532db5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.614821] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc6dfe29-b94d-4727-8ee5-65f7d9f11cdc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.630516] env[61857]: DEBUG nova.compute.provider_tree [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.723426] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950885, 'name': Rename_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.874784] env[61857]: DEBUG nova.network.neutron [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 790.951149] env[61857]: DEBUG oslo_concurrency.lockutils [req-68ba89b5-5435-4944-908d-165c35c21df3 req-8d238cb7-6e2d-4c90-afa6-3c55e33e336f service nova] Releasing lock "refresh_cache-36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.985546] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950886, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062127} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.986625] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.987446] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e17ad9b-716a-46ac-a2cd-49e8c9c177cf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.013403] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 30d905ed-831d-44ae-807c-062de9a7e9fb/30d905ed-831d-44ae-807c-062de9a7e9fb.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 791.016340] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b3454c5-1c6e-4cc9-aac9-937c25933eb7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.038153] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 791.038153] env[61857]: value = "task-950887" [ 791.038153] env[61857]: _type = "Task" [ 791.038153] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.046863] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950887, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.134253] env[61857]: DEBUG nova.scheduler.client.report [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.223238] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950885, 'name': Rename_Task, 'duration_secs': 0.96448} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.223520] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 791.223766] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2acdc7b9-2a45-4cad-8afa-774b955445c0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.229647] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 791.229647] env[61857]: value = "task-950888" [ 791.229647] env[61857]: _type = "Task" [ 791.229647] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.238754] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950888, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.426097] env[61857]: DEBUG nova.network.neutron [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updating instance_info_cache with network_info: [{"id": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "address": "fa:16:3e:40:07:f3", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25a85668-66", "ovs_interfaceid": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.548178] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950887, 'name': ReconfigVM_Task, 'duration_secs': 0.281718} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.549478] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 30d905ed-831d-44ae-807c-062de9a7e9fb/30d905ed-831d-44ae-807c-062de9a7e9fb.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.550996] env[61857]: DEBUG nova.compute.manager [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Received event network-vif-plugged-25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 791.551215] env[61857]: DEBUG oslo_concurrency.lockutils [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] Acquiring lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.551450] env[61857]: DEBUG oslo_concurrency.lockutils [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] Lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.551615] env[61857]: DEBUG oslo_concurrency.lockutils [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] Lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.551784] env[61857]: DEBUG nova.compute.manager [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] No waiting events found dispatching network-vif-plugged-25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 791.551948] env[61857]: WARNING nova.compute.manager [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Received unexpected event network-vif-plugged-25a85668-66ff-4cf6-8cef-68b4d1a2e828 for instance with vm_state building and task_state spawning. [ 791.552122] env[61857]: DEBUG nova.compute.manager [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Received event network-changed-25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 791.552280] env[61857]: DEBUG nova.compute.manager [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Refreshing instance network info cache due to event network-changed-25a85668-66ff-4cf6-8cef-68b4d1a2e828. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 791.552450] env[61857]: DEBUG oslo_concurrency.lockutils [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] Acquiring lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.552668] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8979fcb8-a4e5-493b-b02e-f735d9049485 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.559773] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 791.559773] env[61857]: value = "task-950889" [ 791.559773] env[61857]: _type = "Task" [ 791.559773] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.568078] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950889, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.643607] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.643607] env[61857]: DEBUG nova.compute.manager [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 791.645919] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.387s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.647538] env[61857]: INFO nova.compute.claims [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.739503] env[61857]: DEBUG oslo_vmware.api [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950888, 'name': PowerOnVM_Task, 'duration_secs': 0.44339} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.739816] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 791.739955] env[61857]: INFO nova.compute.manager [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Took 9.68 seconds to spawn the instance on the hypervisor. [ 791.740180] env[61857]: DEBUG nova.compute.manager [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.740893] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6864447-01be-4fc4-b134-9d2a7b68f446 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.929764] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Releasing lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.930151] env[61857]: DEBUG nova.compute.manager [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Instance network_info: |[{"id": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "address": "fa:16:3e:40:07:f3", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25a85668-66", "ovs_interfaceid": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 791.930518] env[61857]: DEBUG oslo_concurrency.lockutils [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] Acquired lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.930720] env[61857]: DEBUG nova.network.neutron [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Refreshing network info cache for port 25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 791.932054] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:07:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afd3feb3-ffcc-4499-a2c2-eb6a48aefde9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25a85668-66ff-4cf6-8cef-68b4d1a2e828', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 791.939915] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Creating folder: Project (2b51cf69f1c947c89a3969daba09361b). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 791.940649] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b28a24a6-8faa-4321-856d-5195662e71bd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.951634] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Created folder: Project (2b51cf69f1c947c89a3969daba09361b) in parent group-v214027. [ 791.951830] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Creating folder: Instances. Parent ref: group-v214092. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 791.952141] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a44c9cc-092d-4d6f-952b-84640b26e978 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.964937] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Created folder: Instances in parent group-v214092. [ 791.965247] env[61857]: DEBUG oslo.service.loopingcall [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.965478] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 791.965661] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1160cfc-2f26-4f0b-8c82-b924a92271f1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.986173] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 791.986173] env[61857]: value = "task-950892" [ 791.986173] env[61857]: _type = "Task" [ 791.986173] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.994127] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950892, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.069434] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950889, 'name': Rename_Task, 'duration_secs': 0.149563} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.069703] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 792.069939] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc865db3-5b03-48ff-b850-e80b9a02bce1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.076151] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 792.076151] env[61857]: value = "task-950893" [ 792.076151] env[61857]: _type = "Task" [ 792.076151] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.083745] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950893, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.152541] env[61857]: DEBUG nova.compute.utils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 792.157991] env[61857]: DEBUG nova.compute.manager [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 792.157991] env[61857]: DEBUG nova.network.neutron [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 792.257889] env[61857]: INFO nova.compute.manager [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Took 38.10 seconds to build instance. [ 792.261140] env[61857]: DEBUG nova.policy [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1882adc2a7e418eb33f3c3c7ba6af20', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80bf659441b343ab987f649d38d2bc9d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 792.501089] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950892, 'name': CreateVM_Task, 'duration_secs': 0.401873} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.505762] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 792.507818] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.507998] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.511905] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 792.511905] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2950588-b33e-45c8-8d8b-b6a379fe1623 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.519491] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 792.519491] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52cf164c-dd3a-5d72-6910-3972bcb5adf8" [ 792.519491] env[61857]: _type = "Task" [ 792.519491] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.534195] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52cf164c-dd3a-5d72-6910-3972bcb5adf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.593599] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950893, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.659130] env[61857]: DEBUG nova.compute.manager [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 792.759803] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cdfe76c1-8419-402e-b2d2-0ea26771c088 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "e91e99b9-1cd3-4345-af09-f14af4df1214" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.694s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.774816] env[61857]: DEBUG nova.network.neutron [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updated VIF entry in instance network info cache for port 25a85668-66ff-4cf6-8cef-68b4d1a2e828. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 792.775206] env[61857]: DEBUG nova.network.neutron [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updating instance_info_cache with network_info: [{"id": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "address": "fa:16:3e:40:07:f3", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25a85668-66", "ovs_interfaceid": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.016277] env[61857]: DEBUG nova.network.neutron [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Successfully created port: 467ad7db-bcf7-4b1c-8fa3-a3d03af32d88 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 793.031956] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52cf164c-dd3a-5d72-6910-3972bcb5adf8, 'name': SearchDatastore_Task, 'duration_secs': 0.010696} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.032296] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.032680] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.032927] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.033091] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.033281] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.033548] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4530635f-423c-4398-9162-1ed8eb8dc671 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.045461] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.045640] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 793.046455] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9ad5faf-b696-4374-9efc-67f8ad06fb8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.052318] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 793.052318] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]528533cd-9a26-5f40-83a9-3d3b2f7fd9d4" [ 793.052318] env[61857]: _type = "Task" [ 793.052318] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.067041] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528533cd-9a26-5f40-83a9-3d3b2f7fd9d4, 'name': SearchDatastore_Task, 'duration_secs': 0.009575} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.067950] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51569fec-b537-42cc-8a5b-9494041b8f52 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.073042] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 793.073042] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52bdac1c-9082-cec9-4751-ac1028439b41" [ 793.073042] env[61857]: _type = "Task" [ 793.073042] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.087777] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52bdac1c-9082-cec9-4751-ac1028439b41, 'name': SearchDatastore_Task, 'duration_secs': 0.008977} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.088361] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.088725] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 24d3d515-e03a-4b4e-bb8e-bc18537125ac/24d3d515-e03a-4b4e-bb8e-bc18537125ac.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 793.089267] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-084cafd9-ded8-4811-bf8c-0ad63118b043 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.094348] env[61857]: DEBUG oslo_vmware.api [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-950893, 'name': PowerOnVM_Task, 'duration_secs': 0.533137} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.094904] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 793.095190] env[61857]: INFO nova.compute.manager [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Took 8.36 seconds to spawn the instance on the hypervisor. [ 793.095686] env[61857]: DEBUG nova.compute.manager [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 793.098975] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15bf4435-e778-445a-9808-0791a84bf846 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.103348] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 793.103348] env[61857]: value = "task-950894" [ 793.103348] env[61857]: _type = "Task" [ 793.103348] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.114958] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.226131] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b88fa35-d3a9-456a-9243-11fa2fc5d87e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.241272] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9bffdf-54fb-4914-9787-df048935ca1c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.272982] env[61857]: DEBUG nova.compute.manager [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 793.276573] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69343c17-076f-4064-859e-fa964fe3642c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.279560] env[61857]: DEBUG oslo_concurrency.lockutils [req-4c03f39d-2872-4a34-8ae6-351c94ee0a3a req-b99c6ba4-5089-4ed2-97c9-9266428869c6 service nova] Releasing lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.285774] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48dbbc5-6739-4a3c-b3be-571f839aa707 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.309333] env[61857]: INFO nova.compute.manager [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Rescuing [ 793.309624] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "refresh_cache-e91e99b9-1cd3-4345-af09-f14af4df1214" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.309775] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "refresh_cache-e91e99b9-1cd3-4345-af09-f14af4df1214" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.310026] env[61857]: DEBUG nova.network.neutron [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 793.311671] env[61857]: DEBUG nova.compute.provider_tree [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.616729] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950894, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.628203] env[61857]: INFO nova.compute.manager [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Took 38.39 seconds to build instance. [ 793.674133] env[61857]: DEBUG nova.compute.manager [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 793.718294] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.718984] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.718984] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.719126] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.719235] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.719385] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.719595] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.719753] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.719923] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.720252] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.720641] env[61857]: DEBUG nova.virt.hardware [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.721842] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729f71e7-1950-402a-a744-3a3fc5d605dd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.731347] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d408a7-ab89-4021-a5d7-41d94902ec7f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.816725] env[61857]: DEBUG nova.scheduler.client.report [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.826274] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.092483] env[61857]: DEBUG nova.network.neutron [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Updating instance_info_cache with network_info: [{"id": "b2d821dc-4eb3-4cb4-962e-7b235495b9be", "address": "fa:16:3e:60:87:33", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2d821dc-4e", "ovs_interfaceid": "b2d821dc-4eb3-4cb4-962e-7b235495b9be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.115878] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950894, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521945} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.116180] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 24d3d515-e03a-4b4e-bb8e-bc18537125ac/24d3d515-e03a-4b4e-bb8e-bc18537125ac.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 794.116416] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 794.116662] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c37c6e04-6b9a-456d-9b2c-cd90bdc31105 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.123367] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 794.123367] env[61857]: value = "task-950895" [ 794.123367] env[61857]: _type = "Task" [ 794.123367] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.132784] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d228cfc-9bdf-4bd5-9522-3f5b4d7dfed2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.187s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.133073] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950895, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.324050] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.677s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.324050] env[61857]: DEBUG nova.compute.manager [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 794.326465] env[61857]: DEBUG oslo_concurrency.lockutils [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 30.778s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.326658] env[61857]: DEBUG nova.objects.instance [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61857) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 794.595625] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "refresh_cache-e91e99b9-1cd3-4345-af09-f14af4df1214" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.636495] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950895, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062375} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.636846] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.638586] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e418148-631b-453a-ad45-e07b41925798 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.641503] env[61857]: DEBUG nova.compute.manager [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.669978] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 24d3d515-e03a-4b4e-bb8e-bc18537125ac/24d3d515-e03a-4b4e-bb8e-bc18537125ac.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.671386] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe01d74c-1456-4073-96d7-178f7ed7fe3d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.692383] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 794.692383] env[61857]: value = "task-950896" [ 794.692383] env[61857]: _type = "Task" [ 794.692383] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.704811] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950896, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.830956] env[61857]: DEBUG nova.compute.utils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.836211] env[61857]: DEBUG nova.compute.manager [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 794.836617] env[61857]: DEBUG nova.network.neutron [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 794.997554] env[61857]: DEBUG nova.policy [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1882adc2a7e418eb33f3c3c7ba6af20', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80bf659441b343ab987f649d38d2bc9d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 795.192734] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.208699] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 795.208813] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60a508aa-fde1-4642-be89-4ddee92ce8e3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.219814] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.222468] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 795.222468] env[61857]: value = "task-950897" [ 795.222468] env[61857]: _type = "Task" [ 795.222468] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.234645] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950897, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.344370] env[61857]: DEBUG nova.compute.manager [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 795.349616] env[61857]: DEBUG oslo_concurrency.lockutils [None req-81280444-f0e6-4610-89dd-960dc4c14016 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.023s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.350420] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.504s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.350651] env[61857]: DEBUG nova.objects.instance [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lazy-loading 'resources' on Instance uuid 7a800abe-ea0a-4077-a5da-dd60eab917e3 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 795.590049] env[61857]: DEBUG nova.network.neutron [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Successfully created port: a930076b-3950-47c6-a108-dc309a22bda4 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 795.706980] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950896, 'name': ReconfigVM_Task, 'duration_secs': 0.623423} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.707128] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 24d3d515-e03a-4b4e-bb8e-bc18537125ac/24d3d515-e03a-4b4e-bb8e-bc18537125ac.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.707825] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8d86b661-d416-4b06-948c-71d6afa9e179 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.715392] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 795.715392] env[61857]: value = "task-950898" [ 795.715392] env[61857]: _type = "Task" [ 795.715392] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.730028] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950898, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.739314] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950897, 'name': PowerOffVM_Task, 'duration_secs': 0.194617} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.739712] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 795.740764] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62381f53-8e94-4090-958d-2fdd5006771b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.761565] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f498d72d-fd02-4f29-a950-56c0284a272d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.829168] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 795.829871] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a260805-0ec3-41d1-afd2-bd66e2a1fc7d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.836935] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 795.836935] env[61857]: value = "task-950899" [ 795.836935] env[61857]: _type = "Task" [ 795.836935] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.847489] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] VM already powered off {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 795.847489] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 795.847489] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.847489] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.847690] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 795.847690] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8fa66f4e-f85c-433d-8cc2-50141965cbed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.857664] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 795.857664] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 795.860713] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cda5b355-7b2c-4704-8769-d7246249636a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.870498] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 795.870498] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52d0c826-6bc6-7ef4-ce50-d505b29465eb" [ 795.870498] env[61857]: _type = "Task" [ 795.870498] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.879467] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52d0c826-6bc6-7ef4-ce50-d505b29465eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.904482] env[61857]: DEBUG nova.compute.manager [req-1172d6e2-593e-4e13-b625-0718a61cdfb3 req-5eca41f2-f2a6-4cea-be0a-5b6ee7b5e24e service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Received event network-vif-plugged-467ad7db-bcf7-4b1c-8fa3-a3d03af32d88 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 795.904482] env[61857]: DEBUG oslo_concurrency.lockutils [req-1172d6e2-593e-4e13-b625-0718a61cdfb3 req-5eca41f2-f2a6-4cea-be0a-5b6ee7b5e24e service nova] Acquiring lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.904482] env[61857]: DEBUG oslo_concurrency.lockutils [req-1172d6e2-593e-4e13-b625-0718a61cdfb3 req-5eca41f2-f2a6-4cea-be0a-5b6ee7b5e24e service nova] Lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.904482] env[61857]: DEBUG oslo_concurrency.lockutils [req-1172d6e2-593e-4e13-b625-0718a61cdfb3 req-5eca41f2-f2a6-4cea-be0a-5b6ee7b5e24e service nova] Lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.904482] env[61857]: DEBUG nova.compute.manager [req-1172d6e2-593e-4e13-b625-0718a61cdfb3 req-5eca41f2-f2a6-4cea-be0a-5b6ee7b5e24e service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] No waiting events found dispatching network-vif-plugged-467ad7db-bcf7-4b1c-8fa3-a3d03af32d88 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 795.904640] env[61857]: WARNING nova.compute.manager [req-1172d6e2-593e-4e13-b625-0718a61cdfb3 req-5eca41f2-f2a6-4cea-be0a-5b6ee7b5e24e service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Received unexpected event network-vif-plugged-467ad7db-bcf7-4b1c-8fa3-a3d03af32d88 for instance with vm_state building and task_state spawning. [ 796.062240] env[61857]: DEBUG nova.network.neutron [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Successfully updated port: 467ad7db-bcf7-4b1c-8fa3-a3d03af32d88 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.216596] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquiring lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.217088] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.217313] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquiring lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.217505] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.217676] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.225755] env[61857]: INFO nova.compute.manager [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Terminating instance [ 796.227803] env[61857]: DEBUG nova.compute.manager [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 796.228947] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 796.228947] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d70eb24-3aa7-4be4-8dc4-1ecb25365ea4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.236616] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950898, 'name': Rename_Task, 'duration_secs': 0.254719} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.236852] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 796.239356] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8c55b14-ba75-4ae5-91c1-2f438a2d7f3c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.242885] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 796.243440] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e84cabbd-1455-4ed5-9834-dfce57ac8784 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.247230] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 796.247230] env[61857]: value = "task-950900" [ 796.247230] env[61857]: _type = "Task" [ 796.247230] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.253506] env[61857]: DEBUG oslo_vmware.api [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for the task: (returnval){ [ 796.253506] env[61857]: value = "task-950901" [ 796.253506] env[61857]: _type = "Task" [ 796.253506] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.259574] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950900, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.264915] env[61857]: DEBUG oslo_vmware.api [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950901, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.364850] env[61857]: DEBUG nova.compute.manager [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 796.381352] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52d0c826-6bc6-7ef4-ce50-d505b29465eb, 'name': SearchDatastore_Task, 'duration_secs': 0.018584} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.383076] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1039a52f-cba1-4dd9-9265-7e08548e7c7a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.389182] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 796.389182] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]522476e4-55d8-2bc3-43e9-fa4815ad5914" [ 796.389182] env[61857]: _type = "Task" [ 796.389182] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.397939] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522476e4-55d8-2bc3-43e9-fa4815ad5914, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.423224] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 796.423224] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 796.423224] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 796.424420] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 796.424420] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 796.424420] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 796.424420] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 796.424420] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 796.424717] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 796.424717] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 796.424717] env[61857]: DEBUG nova.virt.hardware [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.424717] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75abac5b-f0dc-4ccc-9007-837ada8aa1aa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.431368] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db80e6d-da05-4a9e-9550-9fff27cddef5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.501086] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484de82b-a636-4069-b934-fcae5e5941f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.508710] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7d1e82-7931-4601-b440-d8b084b7c21a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.545939] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33ffa1c-ef58-4a67-996c-54292cb209b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.555702] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5467abe7-2927-40e0-aa2b-3a8b763be0d0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.571382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "refresh_cache-42d38f5e-628a-4030-85e1-3ec0595cf3c8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.571542] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquired lock "refresh_cache-42d38f5e-628a-4030-85e1-3ec0595cf3c8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.571708] env[61857]: DEBUG nova.network.neutron [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.573482] env[61857]: DEBUG nova.compute.provider_tree [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 796.669582] env[61857]: DEBUG nova.compute.manager [req-6303f3e5-eaec-4f02-9272-84c698b85b34 req-415c3884-6e78-408f-aced-d42f7c49078a service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Received event network-changed-467ad7db-bcf7-4b1c-8fa3-a3d03af32d88 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 796.669582] env[61857]: DEBUG nova.compute.manager [req-6303f3e5-eaec-4f02-9272-84c698b85b34 req-415c3884-6e78-408f-aced-d42f7c49078a service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Refreshing instance network info cache due to event network-changed-467ad7db-bcf7-4b1c-8fa3-a3d03af32d88. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 796.669582] env[61857]: DEBUG oslo_concurrency.lockutils [req-6303f3e5-eaec-4f02-9272-84c698b85b34 req-415c3884-6e78-408f-aced-d42f7c49078a service nova] Acquiring lock "refresh_cache-42d38f5e-628a-4030-85e1-3ec0595cf3c8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.758204] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950900, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.771234] env[61857]: DEBUG oslo_vmware.api [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950901, 'name': PowerOffVM_Task, 'duration_secs': 0.183491} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.771234] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 796.771234] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 796.771234] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f477887-c897-4979-86b6-65ba18ca5a86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.900103] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522476e4-55d8-2bc3-43e9-fa4815ad5914, 'name': SearchDatastore_Task, 'duration_secs': 0.011612} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.900503] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.900735] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] e91e99b9-1cd3-4345-af09-f14af4df1214/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk. {{(pid=61857) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 796.901013] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9006e372-4e00-405b-aebb-5386b503e1bd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.908632] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 796.908632] env[61857]: value = "task-950903" [ 796.908632] env[61857]: _type = "Task" [ 796.908632] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.916538] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950903, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.098164] env[61857]: ERROR nova.scheduler.client.report [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] [req-4c8de3d5-92ca-49d8-969f-3d33491796fe] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2d5860d0-8574-4e55-9ced-20e66f0314c2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4c8de3d5-92ca-49d8-969f-3d33491796fe"}]} [ 797.116326] env[61857]: DEBUG nova.scheduler.client.report [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Refreshing inventories for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 797.130276] env[61857]: DEBUG nova.network.neutron [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.132992] env[61857]: DEBUG nova.scheduler.client.report [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Updating ProviderTree inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 797.133215] env[61857]: DEBUG nova.compute.provider_tree [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 797.147057] env[61857]: DEBUG nova.scheduler.client.report [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Refreshing aggregate associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, aggregates: None {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 797.176608] env[61857]: DEBUG nova.scheduler.client.report [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Refreshing trait associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 797.258014] env[61857]: DEBUG oslo_vmware.api [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-950900, 'name': PowerOnVM_Task, 'duration_secs': 0.75512} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.260828] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 797.261053] env[61857]: INFO nova.compute.manager [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Took 8.38 seconds to spawn the instance on the hypervisor. [ 797.261291] env[61857]: DEBUG nova.compute.manager [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 797.262710] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0a68a9-c92a-4e0f-9ee7-3af1f4e9c59d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.322674] env[61857]: DEBUG nova.network.neutron [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Updating instance_info_cache with network_info: [{"id": "467ad7db-bcf7-4b1c-8fa3-a3d03af32d88", "address": "fa:16:3e:39:9a:89", "network": {"id": "5325646a-cd92-4840-9d38-5266e2235ce4", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fabda895e0be4fd3b4a2ed5c7cda3a13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d054505-89d3-49c5-8b38-5da917a42c49", "external-id": "nsx-vlan-transportzone-888", "segmentation_id": 888, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap467ad7db-bc", "ovs_interfaceid": "467ad7db-bcf7-4b1c-8fa3-a3d03af32d88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.344178] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 797.344510] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 797.344718] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Deleting the datastore file [datastore2] 2c187f64-8a36-4dfd-94e3-8ea944dbac24 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.347989] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fce5153d-b2e9-4e99-8d9f-994e28d8069c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.355696] env[61857]: DEBUG oslo_vmware.api [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for the task: (returnval){ [ 797.355696] env[61857]: value = "task-950904" [ 797.355696] env[61857]: _type = "Task" [ 797.355696] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.364682] env[61857]: DEBUG oslo_vmware.api [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.403544] env[61857]: DEBUG nova.compute.manager [req-3d9348bb-6ea0-4a65-ba1a-2851ba9d4ec2 req-6274d9ea-c45d-42c7-906d-b748732d2c5c service nova] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Received event network-vif-plugged-a930076b-3950-47c6-a108-dc309a22bda4 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 797.403758] env[61857]: DEBUG oslo_concurrency.lockutils [req-3d9348bb-6ea0-4a65-ba1a-2851ba9d4ec2 req-6274d9ea-c45d-42c7-906d-b748732d2c5c service nova] Acquiring lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.404594] env[61857]: DEBUG oslo_concurrency.lockutils [req-3d9348bb-6ea0-4a65-ba1a-2851ba9d4ec2 req-6274d9ea-c45d-42c7-906d-b748732d2c5c service nova] Lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.404594] env[61857]: DEBUG oslo_concurrency.lockutils [req-3d9348bb-6ea0-4a65-ba1a-2851ba9d4ec2 req-6274d9ea-c45d-42c7-906d-b748732d2c5c service nova] Lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.404594] env[61857]: DEBUG nova.compute.manager [req-3d9348bb-6ea0-4a65-ba1a-2851ba9d4ec2 req-6274d9ea-c45d-42c7-906d-b748732d2c5c service nova] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] No waiting events found dispatching network-vif-plugged-a930076b-3950-47c6-a108-dc309a22bda4 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 797.404594] env[61857]: WARNING nova.compute.manager [req-3d9348bb-6ea0-4a65-ba1a-2851ba9d4ec2 req-6274d9ea-c45d-42c7-906d-b748732d2c5c service nova] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Received unexpected event network-vif-plugged-a930076b-3950-47c6-a108-dc309a22bda4 for instance with vm_state building and task_state spawning. [ 797.421874] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950903, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.496675] env[61857]: DEBUG nova.compute.manager [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Stashing vm_state: active {{(pid=61857) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 797.612705] env[61857]: DEBUG nova.network.neutron [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Successfully updated port: a930076b-3950-47c6-a108-dc309a22bda4 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 797.740661] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ac0120-d4f4-4d13-b470-a8d6d8683a77 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.752291] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5a85e7-ed07-4903-bcdf-09e6ae5e9ce7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.795438] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e4a950-bdf6-4b64-8435-756fe4c7ce14 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.798208] env[61857]: INFO nova.compute.manager [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Took 42.35 seconds to build instance. [ 797.805942] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5762eb-6fc6-4c50-8c3e-281199ccf68c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.821021] env[61857]: DEBUG nova.compute.provider_tree [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 797.826940] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Releasing lock "refresh_cache-42d38f5e-628a-4030-85e1-3ec0595cf3c8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.830108] env[61857]: DEBUG nova.compute.manager [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Instance network_info: |[{"id": "467ad7db-bcf7-4b1c-8fa3-a3d03af32d88", "address": "fa:16:3e:39:9a:89", "network": {"id": "5325646a-cd92-4840-9d38-5266e2235ce4", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fabda895e0be4fd3b4a2ed5c7cda3a13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d054505-89d3-49c5-8b38-5da917a42c49", "external-id": "nsx-vlan-transportzone-888", "segmentation_id": 888, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap467ad7db-bc", "ovs_interfaceid": "467ad7db-bcf7-4b1c-8fa3-a3d03af32d88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 797.830108] env[61857]: DEBUG oslo_concurrency.lockutils [req-6303f3e5-eaec-4f02-9272-84c698b85b34 req-415c3884-6e78-408f-aced-d42f7c49078a service nova] Acquired lock "refresh_cache-42d38f5e-628a-4030-85e1-3ec0595cf3c8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.830485] env[61857]: DEBUG nova.network.neutron [req-6303f3e5-eaec-4f02-9272-84c698b85b34 req-415c3884-6e78-408f-aced-d42f7c49078a service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Refreshing network info cache for port 467ad7db-bcf7-4b1c-8fa3-a3d03af32d88 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 797.830485] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:9a:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d054505-89d3-49c5-8b38-5da917a42c49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '467ad7db-bcf7-4b1c-8fa3-a3d03af32d88', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.838062] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Creating folder: Project (80bf659441b343ab987f649d38d2bc9d). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 797.838874] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c75dd12-ead2-412d-baf3-f09878a140fe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.850836] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Created folder: Project (80bf659441b343ab987f649d38d2bc9d) in parent group-v214027. [ 797.851061] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Creating folder: Instances. Parent ref: group-v214095. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 797.851314] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6263c1a1-1457-4485-957d-c57d38f874ae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.864652] env[61857]: DEBUG oslo_vmware.api [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.865904] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Created folder: Instances in parent group-v214095. [ 797.866154] env[61857]: DEBUG oslo.service.loopingcall [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.866977] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 797.866977] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-efdddc54-fda8-40a1-b329-9f1ed0db7b62 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.889101] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.889101] env[61857]: value = "task-950907" [ 797.889101] env[61857]: _type = "Task" [ 797.889101] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.897471] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950907, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.919368] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950903, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.994192} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.919684] env[61857]: INFO nova.virt.vmwareapi.ds_util [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] e91e99b9-1cd3-4345-af09-f14af4df1214/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk. [ 797.920490] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974050eb-019c-4920-bf18-a51d30005ded {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.946381] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] e91e99b9-1cd3-4345-af09-f14af4df1214/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.946709] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa310931-7487-4f49-9439-6928580c6720 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.964612] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 797.964612] env[61857]: value = "task-950908" [ 797.964612] env[61857]: _type = "Task" [ 797.964612] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.973429] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950908, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.040123] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.115495] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "refresh_cache-6c4c1276-ae80-4f37-9ef9-4872dd415d24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.115679] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquired lock "refresh_cache-6c4c1276-ae80-4f37-9ef9-4872dd415d24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.115844] env[61857]: DEBUG nova.network.neutron [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 798.301705] env[61857]: DEBUG oslo_concurrency.lockutils [None req-03fa06bf-c855-48b9-a7fa-bc57686c73da tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.852s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.367627] env[61857]: DEBUG oslo_vmware.api [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Task: {'id': task-950904, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.87931} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.368067] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 798.368262] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 798.368466] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 798.368661] env[61857]: INFO nova.compute.manager [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Took 2.14 seconds to destroy the instance on the hypervisor. [ 798.368871] env[61857]: DEBUG oslo.service.loopingcall [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.369073] env[61857]: DEBUG nova.compute.manager [-] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.369191] env[61857]: DEBUG nova.network.neutron [-] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 798.379727] env[61857]: DEBUG nova.scheduler.client.report [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Updated inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with generation 84 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 798.379727] env[61857]: DEBUG nova.compute.provider_tree [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Updating resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 generation from 84 to 85 during operation: update_inventory {{(pid=61857) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 798.379727] env[61857]: DEBUG nova.compute.provider_tree [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 798.398878] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950907, 'name': CreateVM_Task, 'duration_secs': 0.351346} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.399662] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 798.400392] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.400567] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.400905] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.401432] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7a43161-0f92-4541-89ad-f596b3b1a78e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.408057] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 798.408057] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52c2165b-bffd-e82b-fbb6-31737c71fdd5" [ 798.408057] env[61857]: _type = "Task" [ 798.408057] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.416369] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c2165b-bffd-e82b-fbb6-31737c71fdd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.482398] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950908, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.684328] env[61857]: DEBUG nova.network.neutron [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 798.793245] env[61857]: DEBUG nova.network.neutron [req-6303f3e5-eaec-4f02-9272-84c698b85b34 req-415c3884-6e78-408f-aced-d42f7c49078a service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Updated VIF entry in instance network info cache for port 467ad7db-bcf7-4b1c-8fa3-a3d03af32d88. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 798.793606] env[61857]: DEBUG nova.network.neutron [req-6303f3e5-eaec-4f02-9272-84c698b85b34 req-415c3884-6e78-408f-aced-d42f7c49078a service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Updating instance_info_cache with network_info: [{"id": "467ad7db-bcf7-4b1c-8fa3-a3d03af32d88", "address": "fa:16:3e:39:9a:89", "network": {"id": "5325646a-cd92-4840-9d38-5266e2235ce4", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fabda895e0be4fd3b4a2ed5c7cda3a13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d054505-89d3-49c5-8b38-5da917a42c49", "external-id": "nsx-vlan-transportzone-888", "segmentation_id": 888, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap467ad7db-bc", "ovs_interfaceid": "467ad7db-bcf7-4b1c-8fa3-a3d03af32d88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.805558] env[61857]: DEBUG nova.compute.manager [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 798.888691] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.535s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.891534] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.165s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.892274] env[61857]: DEBUG nova.objects.instance [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lazy-loading 'resources' on Instance uuid 3cacef7b-dd24-4d33-9500-bbff03e342bb {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 798.920213] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c2165b-bffd-e82b-fbb6-31737c71fdd5, 'name': SearchDatastore_Task, 'duration_secs': 0.032105} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.921239] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.921695] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.922127] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.923249] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.923249] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 798.927035] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7afb7bca-a739-41fb-be24-07a288650ee9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.936352] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 798.936538] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 798.937332] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbd14165-ca74-41d7-852d-37f5559880a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.943280] env[61857]: INFO nova.scheduler.client.report [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Deleted allocations for instance 7a800abe-ea0a-4077-a5da-dd60eab917e3 [ 798.952033] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 798.952033] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525f15aa-b02c-ee0f-53ae-6224738a131a" [ 798.952033] env[61857]: _type = "Task" [ 798.952033] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.961258] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525f15aa-b02c-ee0f-53ae-6224738a131a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.963410] env[61857]: DEBUG nova.network.neutron [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Updating instance_info_cache with network_info: [{"id": "a930076b-3950-47c6-a108-dc309a22bda4", "address": "fa:16:3e:66:55:b0", "network": {"id": "5325646a-cd92-4840-9d38-5266e2235ce4", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.150", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fabda895e0be4fd3b4a2ed5c7cda3a13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d054505-89d3-49c5-8b38-5da917a42c49", "external-id": "nsx-vlan-transportzone-888", "segmentation_id": 888, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa930076b-39", "ovs_interfaceid": "a930076b-3950-47c6-a108-dc309a22bda4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.975354] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950908, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.296860] env[61857]: DEBUG oslo_concurrency.lockutils [req-6303f3e5-eaec-4f02-9272-84c698b85b34 req-415c3884-6e78-408f-aced-d42f7c49078a service nova] Releasing lock "refresh_cache-42d38f5e-628a-4030-85e1-3ec0595cf3c8" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.346802] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.459922] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41ebb07f-efa6-44ed-8361-9e11b161808b tempest-ServerMetadataNegativeTestJSON-986478926 tempest-ServerMetadataNegativeTestJSON-986478926-project-member] Lock "7a800abe-ea0a-4077-a5da-dd60eab917e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.476s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.466247] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Releasing lock "refresh_cache-6c4c1276-ae80-4f37-9ef9-4872dd415d24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.466247] env[61857]: DEBUG nova.compute.manager [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Instance network_info: |[{"id": "a930076b-3950-47c6-a108-dc309a22bda4", "address": "fa:16:3e:66:55:b0", "network": {"id": "5325646a-cd92-4840-9d38-5266e2235ce4", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.150", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fabda895e0be4fd3b4a2ed5c7cda3a13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d054505-89d3-49c5-8b38-5da917a42c49", "external-id": "nsx-vlan-transportzone-888", "segmentation_id": 888, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa930076b-39", "ovs_interfaceid": "a930076b-3950-47c6-a108-dc309a22bda4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 799.466600] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525f15aa-b02c-ee0f-53ae-6224738a131a, 'name': SearchDatastore_Task, 'duration_secs': 0.020603} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.467197] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:55:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d054505-89d3-49c5-8b38-5da917a42c49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a930076b-3950-47c6-a108-dc309a22bda4', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 799.475411] env[61857]: DEBUG oslo.service.loopingcall [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.481382] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 799.481927] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d522ddb1-f035-4b3a-bcc5-b3533eee7cad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.484521] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1fae701-d956-478c-b888-19f33daf24b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.508594] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950908, 'name': ReconfigVM_Task, 'duration_secs': 1.405816} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.508594] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 799.508594] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52487840-b2a8-8dbf-2138-c3d7b6c5db87" [ 799.508594] env[61857]: _type = "Task" [ 799.508594] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.508984] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Reconfigured VM instance instance-00000038 to attach disk [datastore2] e91e99b9-1cd3-4345-af09-f14af4df1214/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.509187] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 799.509187] env[61857]: value = "task-950909" [ 799.509187] env[61857]: _type = "Task" [ 799.509187] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.509880] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0605e1c5-0f86-479a-bda1-05ac0a0cfef0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.553780] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52487840-b2a8-8dbf-2138-c3d7b6c5db87, 'name': SearchDatastore_Task, 'duration_secs': 0.011592} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.559458] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950909, 'name': CreateVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.562590] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.562990] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 42d38f5e-628a-4030-85e1-3ec0595cf3c8/42d38f5e-628a-4030-85e1-3ec0595cf3c8.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 799.563224] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae13afde-6cc9-46b6-82c6-21680d139a05 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.578674] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-963a10cc-828b-49f3-9826-ef7390a8a6b2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.587226] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 799.587226] env[61857]: value = "task-950910" [ 799.587226] env[61857]: _type = "Task" [ 799.587226] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.588552] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 799.588552] env[61857]: value = "task-950911" [ 799.588552] env[61857]: _type = "Task" [ 799.588552] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.599927] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950910, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.605708] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950911, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.668309] env[61857]: DEBUG nova.compute.manager [req-b9d5c3d4-a2c7-4906-a3f3-2f41f0a19cb9 req-1be3e8bb-1daf-46cf-81ab-cda4bd8a1c78 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Received event network-vif-deleted-2104d242-18d6-4e99-b6d0-1dd9efc2778b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 799.668309] env[61857]: INFO nova.compute.manager [req-b9d5c3d4-a2c7-4906-a3f3-2f41f0a19cb9 req-1be3e8bb-1daf-46cf-81ab-cda4bd8a1c78 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Neutron deleted interface 2104d242-18d6-4e99-b6d0-1dd9efc2778b; detaching it from the instance and deleting it from the info cache [ 799.668417] env[61857]: DEBUG nova.network.neutron [req-b9d5c3d4-a2c7-4906-a3f3-2f41f0a19cb9 req-1be3e8bb-1daf-46cf-81ab-cda4bd8a1c78 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.712855] env[61857]: DEBUG nova.network.neutron [-] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.035643] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950909, 'name': CreateVM_Task, 'duration_secs': 0.419642} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.035643] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 800.035643] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.035643] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.035643] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 800.035643] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d28202e2-8b8a-4fec-8ab5-cf75f99a1f70 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.043254] env[61857]: DEBUG nova.compute.manager [req-b19126a1-2c75-4b7a-af7c-c3d12c980b64 req-97e1ba91-792a-4f95-b156-ed78af4152f0 service nova] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Received event network-changed-a930076b-3950-47c6-a108-dc309a22bda4 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 800.043603] env[61857]: DEBUG nova.compute.manager [req-b19126a1-2c75-4b7a-af7c-c3d12c980b64 req-97e1ba91-792a-4f95-b156-ed78af4152f0 service nova] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Refreshing instance network info cache due to event network-changed-a930076b-3950-47c6-a108-dc309a22bda4. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 800.044404] env[61857]: DEBUG oslo_concurrency.lockutils [req-b19126a1-2c75-4b7a-af7c-c3d12c980b64 req-97e1ba91-792a-4f95-b156-ed78af4152f0 service nova] Acquiring lock "refresh_cache-6c4c1276-ae80-4f37-9ef9-4872dd415d24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.044882] env[61857]: DEBUG oslo_concurrency.lockutils [req-b19126a1-2c75-4b7a-af7c-c3d12c980b64 req-97e1ba91-792a-4f95-b156-ed78af4152f0 service nova] Acquired lock "refresh_cache-6c4c1276-ae80-4f37-9ef9-4872dd415d24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.045161] env[61857]: DEBUG nova.network.neutron [req-b19126a1-2c75-4b7a-af7c-c3d12c980b64 req-97e1ba91-792a-4f95-b156-ed78af4152f0 service nova] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Refreshing network info cache for port a930076b-3950-47c6-a108-dc309a22bda4 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.053041] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 800.053041] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525f6b29-a2be-1208-ce76-2f43e8442482" [ 800.053041] env[61857]: _type = "Task" [ 800.053041] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.063364] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525f6b29-a2be-1208-ce76-2f43e8442482, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.069199] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd04832c-43c2-406b-9ea3-a8ef005899eb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.077625] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fe60ac-88f8-4a79-bc67-7ed25c13eedc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.122651] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375d078f-728b-408e-aed2-731da6f57fd8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.130609] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950910, 'name': ReconfigVM_Task, 'duration_secs': 0.506873} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.135473] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 800.135798] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950911, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484628} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.136013] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4c1d7ca-b759-4430-8b61-4cdd13cf1dc9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.138506] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e739a8df-12c3-4188-9e51-3d1954624fd1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.142475] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 42d38f5e-628a-4030-85e1-3ec0595cf3c8/42d38f5e-628a-4030-85e1-3ec0595cf3c8.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 800.142704] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 800.142954] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9dce071b-3a87-4385-8ad0-ae91cedf727d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.155438] env[61857]: DEBUG nova.compute.provider_tree [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 800.158711] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 800.158711] env[61857]: value = "task-950912" [ 800.158711] env[61857]: _type = "Task" [ 800.158711] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.158962] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 800.158962] env[61857]: value = "task-950913" [ 800.158962] env[61857]: _type = "Task" [ 800.158962] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.173022] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950912, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.174557] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950913, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.174653] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc61eb45-6371-448e-b66a-47649d2a6535 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.183194] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb6ec5a-ce60-43aa-ab94-b15b63f28a36 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.216535] env[61857]: DEBUG nova.compute.manager [req-b9d5c3d4-a2c7-4906-a3f3-2f41f0a19cb9 req-1be3e8bb-1daf-46cf-81ab-cda4bd8a1c78 service nova] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Detach interface failed, port_id=2104d242-18d6-4e99-b6d0-1dd9efc2778b, reason: Instance 2c187f64-8a36-4dfd-94e3-8ea944dbac24 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 800.218328] env[61857]: INFO nova.compute.manager [-] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Took 1.85 seconds to deallocate network for instance. [ 800.567625] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525f6b29-a2be-1208-ce76-2f43e8442482, 'name': SearchDatastore_Task, 'duration_secs': 0.010394} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.568018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.568420] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 800.568728] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.568923] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.569159] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 800.569496] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eda30c7a-0154-4352-83ab-f6893c1bf651 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.583220] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 800.583220] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 800.584427] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0866b7eb-7423-45d1-bf73-7a4e260adf73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.595491] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 800.595491] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fbfa52-7d36-d5c8-2ecc-10de2bb5b6ef" [ 800.595491] env[61857]: _type = "Task" [ 800.595491] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.609432] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fbfa52-7d36-d5c8-2ecc-10de2bb5b6ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.678619] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950913, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064831} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.678859] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950912, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.681743] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 800.683255] env[61857]: ERROR nova.scheduler.client.report [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [req-c5870329-d7cf-44c8-85d9-f01e49f66a23] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2d5860d0-8574-4e55-9ced-20e66f0314c2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c5870329-d7cf-44c8-85d9-f01e49f66a23"}]} [ 800.684102] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd01be96-9404-416f-9a92-0c49b0afb3a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.713165] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 42d38f5e-628a-4030-85e1-3ec0595cf3c8/42d38f5e-628a-4030-85e1-3ec0595cf3c8.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.714276] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5854ca4-186d-4cad-bc54-8761481c84c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.729159] env[61857]: DEBUG nova.scheduler.client.report [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Refreshing inventories for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 800.731743] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.738719] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 800.738719] env[61857]: value = "task-950914" [ 800.738719] env[61857]: _type = "Task" [ 800.738719] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.747382] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950914, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.748714] env[61857]: DEBUG nova.scheduler.client.report [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Updating ProviderTree inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 800.748714] env[61857]: DEBUG nova.compute.provider_tree [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 800.773944] env[61857]: DEBUG nova.scheduler.client.report [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Refreshing aggregate associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, aggregates: None {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 800.802442] env[61857]: DEBUG nova.scheduler.client.report [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Refreshing trait associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 800.921288] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "ffa63de7-dd32-4908-92c5-755b3c0799ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.921918] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.968906] env[61857]: DEBUG nova.network.neutron [req-b19126a1-2c75-4b7a-af7c-c3d12c980b64 req-97e1ba91-792a-4f95-b156-ed78af4152f0 service nova] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Updated VIF entry in instance network info cache for port a930076b-3950-47c6-a108-dc309a22bda4. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 800.969318] env[61857]: DEBUG nova.network.neutron [req-b19126a1-2c75-4b7a-af7c-c3d12c980b64 req-97e1ba91-792a-4f95-b156-ed78af4152f0 service nova] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Updating instance_info_cache with network_info: [{"id": "a930076b-3950-47c6-a108-dc309a22bda4", "address": "fa:16:3e:66:55:b0", "network": {"id": "5325646a-cd92-4840-9d38-5266e2235ce4", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.150", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "fabda895e0be4fd3b4a2ed5c7cda3a13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d054505-89d3-49c5-8b38-5da917a42c49", "external-id": "nsx-vlan-transportzone-888", "segmentation_id": 888, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa930076b-39", "ovs_interfaceid": "a930076b-3950-47c6-a108-dc309a22bda4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.108312] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fbfa52-7d36-d5c8-2ecc-10de2bb5b6ef, 'name': SearchDatastore_Task, 'duration_secs': 0.017957} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.109887] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-923e075b-6a24-456d-a93f-679a7397d208 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.118015] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 801.118015] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52633f4f-5006-07cc-ca42-864dfb9cfd47" [ 801.118015] env[61857]: _type = "Task" [ 801.118015] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.130670] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52633f4f-5006-07cc-ca42-864dfb9cfd47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.179284] env[61857]: DEBUG oslo_vmware.api [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950912, 'name': PowerOnVM_Task, 'duration_secs': 0.94818} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.179562] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 801.182466] env[61857]: DEBUG nova.compute.manager [None req-ffac747f-b97c-43c5-b458-5f14c4736ffc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 801.183414] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c98c6cb-5a5f-4904-bee6-8233ab2ab948 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.250138] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950914, 'name': ReconfigVM_Task, 'duration_secs': 0.306323} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.250645] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 42d38f5e-628a-4030-85e1-3ec0595cf3c8/42d38f5e-628a-4030-85e1-3ec0595cf3c8.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 801.251013] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-901b4ed6-8b17-44fa-9f49-2cc54a60209c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.258624] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 801.258624] env[61857]: value = "task-950915" [ 801.258624] env[61857]: _type = "Task" [ 801.258624] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.269820] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950915, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.360366] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c2421b-e87b-465b-bd26-98a8434c2816 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.368525] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1fc825-918d-4a45-9b8d-9cd54525b41b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.406503] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef6d13b-7f81-4768-94ab-90eec12c2461 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.415319] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88abfbe9-4539-4d95-ba43-60a228433357 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.430308] env[61857]: DEBUG nova.compute.provider_tree [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.471972] env[61857]: DEBUG oslo_concurrency.lockutils [req-b19126a1-2c75-4b7a-af7c-c3d12c980b64 req-97e1ba91-792a-4f95-b156-ed78af4152f0 service nova] Releasing lock "refresh_cache-6c4c1276-ae80-4f37-9ef9-4872dd415d24" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.634757] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52633f4f-5006-07cc-ca42-864dfb9cfd47, 'name': SearchDatastore_Task, 'duration_secs': 0.031198} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.635511] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.635511] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 6c4c1276-ae80-4f37-9ef9-4872dd415d24/6c4c1276-ae80-4f37-9ef9-4872dd415d24.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 801.635659] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bed3243b-92f8-4eac-b042-88e6a143cf27 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.642524] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 801.642524] env[61857]: value = "task-950916" [ 801.642524] env[61857]: _type = "Task" [ 801.642524] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.650185] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950916, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.771771] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950915, 'name': Rename_Task, 'duration_secs': 0.340965} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.772082] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 801.772342] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c039121-59a0-4cd4-a887-39b28faab3e9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.778837] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 801.778837] env[61857]: value = "task-950917" [ 801.778837] env[61857]: _type = "Task" [ 801.778837] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.787302] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950917, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.936382] env[61857]: DEBUG nova.scheduler.client.report [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.156215] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950916, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.289783] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950917, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.443466] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.551s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.446039] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 34.105s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.446039] env[61857]: DEBUG nova.objects.instance [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61857) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 802.496830] env[61857]: INFO nova.scheduler.client.report [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Deleted allocations for instance 3cacef7b-dd24-4d33-9500-bbff03e342bb [ 802.655246] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950916, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.626644} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.655246] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 6c4c1276-ae80-4f37-9ef9-4872dd415d24/6c4c1276-ae80-4f37-9ef9-4872dd415d24.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 802.655693] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 802.656109] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c878317e-97d4-4967-aff0-b1514d505b11 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.662776] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 802.662776] env[61857]: value = "task-950918" [ 802.662776] env[61857]: _type = "Task" [ 802.662776] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.675041] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950918, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.790892] env[61857]: DEBUG oslo_vmware.api [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950917, 'name': PowerOnVM_Task, 'duration_secs': 0.609625} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.791272] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 802.791552] env[61857]: INFO nova.compute.manager [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Took 9.12 seconds to spawn the instance on the hypervisor. [ 802.791769] env[61857]: DEBUG nova.compute.manager [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 802.792642] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f33604-50f4-42e7-adf3-24f0dfd4837b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.009139] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f901f431-b920-4ffb-83d1-459761c626d1 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "3cacef7b-dd24-4d33-9500-bbff03e342bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.979s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.173842] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950918, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118589} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.174184] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.174980] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a4ab0b-2dd5-4f90-95cb-5dd1fd3d7c77 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.199253] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 6c4c1276-ae80-4f37-9ef9-4872dd415d24/6c4c1276-ae80-4f37-9ef9-4872dd415d24.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.199617] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60d1a1ef-29e1-4f45-88c9-c1b16215354b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.221231] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 803.221231] env[61857]: value = "task-950919" [ 803.221231] env[61857]: _type = "Task" [ 803.221231] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.231317] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950919, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.315625] env[61857]: INFO nova.compute.manager [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Took 43.29 seconds to build instance. [ 803.464612] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9fa57db7-780d-4ab9-af3f-aae1a10c6e93 tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.465171] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.768s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.465725] env[61857]: DEBUG nova.objects.instance [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lazy-loading 'resources' on Instance uuid e286e6d0-25db-4d86-b9a2-3538bc3f624a {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 803.567856] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquiring lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.569096] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.731788] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950919, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.817774] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e3bca310-3001-4f08-bd9b-dc7bbb11ad64 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.330s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.876177] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "d49008e6-7789-4785-8eea-9b3fa36c4ccb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.876456] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "d49008e6-7789-4785-8eea-9b3fa36c4ccb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.876664] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "d49008e6-7789-4785-8eea-9b3fa36c4ccb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.876889] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "d49008e6-7789-4785-8eea-9b3fa36c4ccb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.877031] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "d49008e6-7789-4785-8eea-9b3fa36c4ccb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.879758] env[61857]: INFO nova.compute.manager [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Terminating instance [ 803.881588] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "refresh_cache-d49008e6-7789-4785-8eea-9b3fa36c4ccb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.881747] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquired lock "refresh_cache-d49008e6-7789-4785-8eea-9b3fa36c4ccb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.881914] env[61857]: DEBUG nova.network.neutron [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.128650] env[61857]: INFO nova.compute.manager [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Rescuing [ 804.128930] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.130244] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.130483] env[61857]: DEBUG nova.network.neutron [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.234966] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950919, 'name': ReconfigVM_Task, 'duration_secs': 0.841931} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.234966] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 6c4c1276-ae80-4f37-9ef9-4872dd415d24/6c4c1276-ae80-4f37-9ef9-4872dd415d24.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.235808] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d1aa2b3-35a5-4c08-a4a8-aca84ed8a583 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.243019] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 804.243019] env[61857]: value = "task-950920" [ 804.243019] env[61857]: _type = "Task" [ 804.243019] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.251542] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950920, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.320703] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 804.404037] env[61857]: DEBUG nova.network.neutron [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 804.445109] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab21d0ce-8f2f-49ef-986f-6b79fdfbb824 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.453374] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14c2aa6-1f6f-4c70-a6e6-0d9933213037 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.458479] env[61857]: DEBUG nova.network.neutron [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.486514] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d87f0f22-ebb8-4688-856f-48da77c4dd6f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.496314] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2313e4d7-7637-4256-b281-64a4ee6cefa5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.512020] env[61857]: DEBUG nova.compute.provider_tree [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 804.752983] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950920, 'name': Rename_Task, 'duration_secs': 0.428941} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.753243] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 804.753532] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5395321-4b6d-4332-bf84-a2b99437cb8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.760065] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 804.760065] env[61857]: value = "task-950921" [ 804.760065] env[61857]: _type = "Task" [ 804.760065] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.769856] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950921, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.843322] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.941841] env[61857]: DEBUG nova.network.neutron [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Updating instance_info_cache with network_info: [{"id": "45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6", "address": "fa:16:3e:05:b6:52", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45483cb0-62", "ovs_interfaceid": "45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.986359] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Releasing lock "refresh_cache-d49008e6-7789-4785-8eea-9b3fa36c4ccb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.986915] env[61857]: DEBUG nova.compute.manager [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 804.987270] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 804.988497] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da493f81-d553-47fe-86a9-4fd57775cfda {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.999906] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 805.001992] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9d7865b-0150-4fde-a5d4-c6c873ff841e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.007260] env[61857]: DEBUG oslo_vmware.api [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 805.007260] env[61857]: value = "task-950922" [ 805.007260] env[61857]: _type = "Task" [ 805.007260] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.024171] env[61857]: DEBUG oslo_vmware.api [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950922, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.052026] env[61857]: ERROR nova.scheduler.client.report [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] [req-f760e4a7-84c7-41b4-a126-ece07b43520a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2d5860d0-8574-4e55-9ced-20e66f0314c2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f760e4a7-84c7-41b4-a126-ece07b43520a"}]} [ 805.074902] env[61857]: DEBUG nova.scheduler.client.report [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Refreshing inventories for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 805.096760] env[61857]: DEBUG nova.scheduler.client.report [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Updating ProviderTree inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 805.096760] env[61857]: DEBUG nova.compute.provider_tree [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 805.122369] env[61857]: DEBUG nova.scheduler.client.report [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Refreshing aggregate associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, aggregates: None {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 805.155081] env[61857]: DEBUG nova.scheduler.client.report [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Refreshing trait associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 805.276393] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950921, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.446886] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.523559] env[61857]: DEBUG oslo_vmware.api [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950922, 'name': PowerOffVM_Task, 'duration_secs': 0.231051} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.523834] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 805.524051] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 805.524277] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-549ffe35-ae29-4e24-bdee-21c27134425c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.550847] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 805.550847] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 805.550847] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Deleting the datastore file [datastore1] d49008e6-7789-4785-8eea-9b3fa36c4ccb {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.550847] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c69d6633-ee55-4615-b7b4-51a4bf06372d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.555548] env[61857]: DEBUG oslo_vmware.api [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for the task: (returnval){ [ 805.555548] env[61857]: value = "task-950924" [ 805.555548] env[61857]: _type = "Task" [ 805.555548] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.567400] env[61857]: DEBUG oslo_vmware.api [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950924, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.723666] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4b5607-b55e-4276-9449-8f227bfb878e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.731897] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824c0a00-4347-432f-82c8-ec7edc2e6a12 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.766105] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c351d8-3b70-4b33-b37b-b6b3e15ec341 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.777448] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881acddf-55d1-444f-a7e7-6d55aecafe4a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.781408] env[61857]: DEBUG oslo_vmware.api [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950921, 'name': PowerOnVM_Task, 'duration_secs': 0.577199} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.781692] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 805.781895] env[61857]: INFO nova.compute.manager [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Took 9.42 seconds to spawn the instance on the hypervisor. [ 805.782082] env[61857]: DEBUG nova.compute.manager [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.783252] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d61af7c-92f5-42a6-9be5-32228f2e51cc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.793666] env[61857]: DEBUG nova.compute.provider_tree [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 805.988594] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 805.988594] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f688805b-e156-483f-8722-d6383e37bccd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.997734] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 805.997734] env[61857]: value = "task-950925" [ 805.997734] env[61857]: _type = "Task" [ 805.997734] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.008938] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950925, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.069344] env[61857]: DEBUG oslo_vmware.api [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950924, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.307842] env[61857]: INFO nova.compute.manager [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Took 43.07 seconds to build instance. [ 806.325392] env[61857]: DEBUG nova.scheduler.client.report [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Updated inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with generation 87 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 806.325655] env[61857]: DEBUG nova.compute.provider_tree [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Updating resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 generation from 87 to 88 during operation: update_inventory {{(pid=61857) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 806.325835] env[61857]: DEBUG nova.compute.provider_tree [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 806.507115] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950925, 'name': PowerOffVM_Task, 'duration_secs': 0.438843} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.507383] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 806.508146] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f9543a-26ce-436a-9d91-7b62662cda1f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.526904] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9744c6c-918c-4d16-8241-46b4abd5356c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.555021] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 806.555021] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df594c2d-b122-46ef-8d5f-d6c3d341b2d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.561338] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 806.561338] env[61857]: value = "task-950926" [ 806.561338] env[61857]: _type = "Task" [ 806.561338] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.567341] env[61857]: DEBUG oslo_vmware.api [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Task: {'id': task-950924, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.73108} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.567760] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 806.567946] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 806.568155] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 806.568333] env[61857]: INFO nova.compute.manager [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Took 1.58 seconds to destroy the instance on the hypervisor. [ 806.568562] env[61857]: DEBUG oslo.service.loopingcall [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.568751] env[61857]: DEBUG nova.compute.manager [-] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.568846] env[61857]: DEBUG nova.network.neutron [-] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 806.573597] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] VM already powered off {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 806.573787] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.574029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.574184] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.574430] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.574583] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-515d76a2-c2f4-4596-9620-b2ee5eb83b28 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.582017] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.582196] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 806.582898] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd2090b0-1861-4b03-a988-700796ac37ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.586809] env[61857]: DEBUG nova.network.neutron [-] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.589316] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 806.589316] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52555849-7256-4f5b-6eaf-40c70e3b9a72" [ 806.589316] env[61857]: _type = "Task" [ 806.589316] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.597466] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52555849-7256-4f5b-6eaf-40c70e3b9a72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.810246] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a05e2fe6-950c-46da-a325-dc0c77c48abe tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.768s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.831120] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.366s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.833523] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.386s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.835292] env[61857]: INFO nova.compute.claims [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.849401] env[61857]: INFO nova.scheduler.client.report [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Deleted allocations for instance e286e6d0-25db-4d86-b9a2-3538bc3f624a [ 807.091119] env[61857]: DEBUG nova.network.neutron [-] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.105619] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52555849-7256-4f5b-6eaf-40c70e3b9a72, 'name': SearchDatastore_Task, 'duration_secs': 0.008318} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.107262] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54ef9458-f9cf-47c6-85cb-0bbb2240c789 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.113943] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 807.113943] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e463c2-d93a-fc28-ef76-68763ac1e6fb" [ 807.113943] env[61857]: _type = "Task" [ 807.113943] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.125665] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e463c2-d93a-fc28-ef76-68763ac1e6fb, 'name': SearchDatastore_Task, 'duration_secs': 0.0092} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.125969] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.126361] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 28bf59ca-4ffe-4005-9a88-da0660ebb48a/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk. {{(pid=61857) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 807.126667] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-277844c9-4405-4495-8844-17680b0dd168 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.134785] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 807.134785] env[61857]: value = "task-950927" [ 807.134785] env[61857]: _type = "Task" [ 807.134785] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.142936] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950927, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.313779] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 807.357583] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62622ce-3f99-43fe-a223-c2a76f0e2baa tempest-ServerShowV254Test-1035324332 tempest-ServerShowV254Test-1035324332-project-member] Lock "e286e6d0-25db-4d86-b9a2-3538bc3f624a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.439s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.598669] env[61857]: INFO nova.compute.manager [-] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Took 1.03 seconds to deallocate network for instance. [ 807.646472] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950927, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463357} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.646936] env[61857]: INFO nova.virt.vmwareapi.ds_util [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 28bf59ca-4ffe-4005-9a88-da0660ebb48a/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk. [ 807.648401] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dae41e3-c238-41c5-b2ef-eb3a26180a38 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.678253] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 28bf59ca-4ffe-4005-9a88-da0660ebb48a/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.678861] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7c54d68-549f-423e-8e79-a8749c6b6cd9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.701077] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 807.701077] env[61857]: value = "task-950928" [ 807.701077] env[61857]: _type = "Task" [ 807.701077] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.708174] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950928, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.839279] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.105861] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.210456] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950928, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.235206] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9330ced8-a0b8-4db4-b0c3-668fb0cd1bb0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.242609] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e2518d-7aa7-4ee9-942d-af806f790671 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.274156] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb57b89-12a1-4045-96a8-690cdc8dc965 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.282569] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0ec4c36-32a8-4ee9-aa9b-cb166b115459 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.296669] env[61857]: DEBUG nova.compute.manager [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 808.297169] env[61857]: DEBUG nova.compute.provider_tree [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.298856] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a830bb9-2eff-4ceb-843b-cd91216c68b9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.711520] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950928, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.802919] env[61857]: DEBUG nova.scheduler.client.report [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.808578] env[61857]: INFO nova.compute.manager [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] instance snapshotting [ 808.811267] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25a8890-0d38-46f6-835b-14cc29ece689 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.831161] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e473cf0e-a88b-41eb-b257-7cc4cd7bfab3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.214825] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950928, 'name': ReconfigVM_Task, 'duration_secs': 1.046467} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.215148] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 28bf59ca-4ffe-4005-9a88-da0660ebb48a/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.215969] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea3f304-d57e-464e-8b82-988f9c034b53 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.244526] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e1a0459-a7d5-4a75-92d2-79ffeeb4bf2e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.260385] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 809.260385] env[61857]: value = "task-950929" [ 809.260385] env[61857]: _type = "Task" [ 809.260385] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.268096] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950929, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.308229] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.308771] env[61857]: DEBUG nova.compute.manager [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.311672] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.660s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.313061] env[61857]: INFO nova.compute.claims [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.342917] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 809.342917] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e89515e3-f312-46e3-880f-01c68e200229 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.351439] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 809.351439] env[61857]: value = "task-950930" [ 809.351439] env[61857]: _type = "Task" [ 809.351439] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.362515] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950930, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.769817] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950929, 'name': ReconfigVM_Task, 'duration_secs': 0.210121} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.770259] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 809.770378] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c725e04-ae3a-4ec2-9e7d-d617da9b7384 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.777278] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 809.777278] env[61857]: value = "task-950931" [ 809.777278] env[61857]: _type = "Task" [ 809.777278] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.784382] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950931, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.820401] env[61857]: DEBUG nova.compute.utils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.821903] env[61857]: DEBUG nova.compute.manager [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 809.822121] env[61857]: DEBUG nova.network.neutron [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 809.862058] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950930, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.870817] env[61857]: DEBUG nova.policy [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9056d344ff614950a60983450a2a40ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5653ed6ab0a948ef89bb03cf7426dd10', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.291020] env[61857]: DEBUG oslo_vmware.api [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950931, 'name': PowerOnVM_Task, 'duration_secs': 0.443227} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.291020] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 810.291869] env[61857]: DEBUG nova.compute.manager [None req-dac8e5d8-2a6a-43ec-83b8-c23b57acf4f4 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.292653] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc6da70-796f-420a-962e-3e7b92e52f86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.325992] env[61857]: DEBUG nova.compute.manager [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.365026] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950930, 'name': CreateSnapshot_Task, 'duration_secs': 0.820175} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.365026] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 810.365869] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd580257-6df4-47b6-935f-a8c7fe7b89e2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.567398] env[61857]: DEBUG nova.network.neutron [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Successfully created port: ea2b08a8-d692-400f-a7a5-124e3a5c14de {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.844887] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274b0152-706c-43b6-be69-233c8d2632b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.852357] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16daf076-ea1b-4641-bbd4-2a7e1b17da73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.891890] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 810.891890] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1fe0d1fa-a00f-4b8f-a608-a83cda709f73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.894885] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca03d723-1eab-4cbc-9030-5911a63511d7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.904990] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33314f3a-88ea-41e6-9b87-bfb001cf23d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.908930] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 810.908930] env[61857]: value = "task-950932" [ 810.908930] env[61857]: _type = "Task" [ 810.908930] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.920226] env[61857]: DEBUG nova.compute.provider_tree [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.928997] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950932, 'name': CloneVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.336517] env[61857]: DEBUG nova.compute.manager [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.361558] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.361805] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.361962] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.362341] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.362341] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.362467] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.362638] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.362801] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.362965] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.363141] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.363322] env[61857]: DEBUG nova.virt.hardware [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.364222] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1d8fd1-f722-4d57-8578-8688e26da312 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.372364] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb1722b-e6bf-4e17-ae3d-3ce4022b790a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.419136] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950932, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.423156] env[61857]: DEBUG nova.scheduler.client.report [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.441796] env[61857]: INFO nova.compute.manager [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Unrescuing [ 811.442052] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.442212] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.442380] env[61857]: DEBUG nova.network.neutron [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.922564] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950932, 'name': CloneVM_Task} progress is 95%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.927514] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.616s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.927996] env[61857]: DEBUG nova.compute.manager [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.930837] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 28.723s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.930999] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.931228] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 811.931585] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.967s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.933033] env[61857]: INFO nova.compute.claims [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.936337] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d827417-129d-4c11-ab10-718570e5956a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.944949] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1a09f8-3359-431b-822b-7eee88e8b5ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.963140] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c60969-b614-40cd-bb28-04b464d0391a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.970514] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49d6352-f982-4cab-83e7-ed9641f60946 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.002805] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180577MB free_disk=40GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 812.002974] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.165711] env[61857]: DEBUG nova.compute.manager [req-0072e819-0ef7-4fa7-b989-2426bddafb7f req-3e7447b1-eece-40d1-8d95-8a0f0d841eda service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Received event network-vif-plugged-ea2b08a8-d692-400f-a7a5-124e3a5c14de {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 812.165711] env[61857]: DEBUG oslo_concurrency.lockutils [req-0072e819-0ef7-4fa7-b989-2426bddafb7f req-3e7447b1-eece-40d1-8d95-8a0f0d841eda service nova] Acquiring lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.165711] env[61857]: DEBUG oslo_concurrency.lockutils [req-0072e819-0ef7-4fa7-b989-2426bddafb7f req-3e7447b1-eece-40d1-8d95-8a0f0d841eda service nova] Lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.165711] env[61857]: DEBUG oslo_concurrency.lockutils [req-0072e819-0ef7-4fa7-b989-2426bddafb7f req-3e7447b1-eece-40d1-8d95-8a0f0d841eda service nova] Lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.165711] env[61857]: DEBUG nova.compute.manager [req-0072e819-0ef7-4fa7-b989-2426bddafb7f req-3e7447b1-eece-40d1-8d95-8a0f0d841eda service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] No waiting events found dispatching network-vif-plugged-ea2b08a8-d692-400f-a7a5-124e3a5c14de {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.165920] env[61857]: WARNING nova.compute.manager [req-0072e819-0ef7-4fa7-b989-2426bddafb7f req-3e7447b1-eece-40d1-8d95-8a0f0d841eda service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Received unexpected event network-vif-plugged-ea2b08a8-d692-400f-a7a5-124e3a5c14de for instance with vm_state building and task_state spawning. [ 812.210983] env[61857]: DEBUG nova.network.neutron [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Updating instance_info_cache with network_info: [{"id": "45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6", "address": "fa:16:3e:05:b6:52", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45483cb0-62", "ovs_interfaceid": "45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.262392] env[61857]: DEBUG nova.network.neutron [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Successfully updated port: ea2b08a8-d692-400f-a7a5-124e3a5c14de {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.426965] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950932, 'name': CloneVM_Task, 'duration_secs': 1.335657} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.427427] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Created linked-clone VM from snapshot [ 812.428428] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06bb7ead-39be-417a-bac7-9b4eac920f18 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.438085] env[61857]: DEBUG nova.compute.utils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.439482] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Uploading image 3ab5d30a-b778-412e-a710-7edebe17a33f {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 812.441985] env[61857]: DEBUG nova.compute.manager [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.442204] env[61857]: DEBUG nova.network.neutron [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 812.466760] env[61857]: DEBUG oslo_vmware.rw_handles [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 812.466760] env[61857]: value = "vm-214100" [ 812.466760] env[61857]: _type = "VirtualMachine" [ 812.466760] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 812.467073] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9bb8ee80-2973-436c-84c7-9b969ece2d69 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.474383] env[61857]: DEBUG oslo_vmware.rw_handles [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lease: (returnval){ [ 812.474383] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52954901-9207-488b-2193-b3ecb998e821" [ 812.474383] env[61857]: _type = "HttpNfcLease" [ 812.474383] env[61857]: } obtained for exporting VM: (result){ [ 812.474383] env[61857]: value = "vm-214100" [ 812.474383] env[61857]: _type = "VirtualMachine" [ 812.474383] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 812.474629] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the lease: (returnval){ [ 812.474629] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52954901-9207-488b-2193-b3ecb998e821" [ 812.474629] env[61857]: _type = "HttpNfcLease" [ 812.474629] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 812.480998] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 812.480998] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52954901-9207-488b-2193-b3ecb998e821" [ 812.480998] env[61857]: _type = "HttpNfcLease" [ 812.480998] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 812.509728] env[61857]: DEBUG nova.policy [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74f5499cbed541deaf1905605e10158f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62bd314daba846cd9ed3dc994dd1034b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.714180] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "refresh_cache-28bf59ca-4ffe-4005-9a88-da0660ebb48a" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.714912] env[61857]: DEBUG nova.objects.instance [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lazy-loading 'flavor' on Instance uuid 28bf59ca-4ffe-4005-9a88-da0660ebb48a {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.771702] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.771702] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.771702] env[61857]: DEBUG nova.network.neutron [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 812.812018] env[61857]: DEBUG nova.network.neutron [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Successfully created port: 7ae4b256-08e6-483a-a76a-1c1bcc44e67d {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.943390] env[61857]: DEBUG nova.compute.manager [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.987551] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 812.987551] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52954901-9207-488b-2193-b3ecb998e821" [ 812.987551] env[61857]: _type = "HttpNfcLease" [ 812.987551] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 812.987867] env[61857]: DEBUG oslo_vmware.rw_handles [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 812.987867] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52954901-9207-488b-2193-b3ecb998e821" [ 812.987867] env[61857]: _type = "HttpNfcLease" [ 812.987867] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 812.988660] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ec8904e-49be-4513-bc32-297c2cb3fc49 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.998145] env[61857]: DEBUG oslo_vmware.rw_handles [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52737eee-4b9a-07b9-31bc-65759d471c5c/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 812.998327] env[61857]: DEBUG oslo_vmware.rw_handles [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52737eee-4b9a-07b9-31bc-65759d471c5c/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 813.199695] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-16c54595-4aa2-4434-8eb1-2960a349fda6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.223375] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1d7b1c-93f8-420d-9441-737cd9c856e0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.248153] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.255369] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d191b37d-e0b5-48df-8819-2d11fa5ea726 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.263023] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 813.263023] env[61857]: value = "task-950934" [ 813.263023] env[61857]: _type = "Task" [ 813.263023] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.268851] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950934, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.326249] env[61857]: DEBUG nova.network.neutron [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.343302] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Acquiring lock "c4883f84-e62b-4b59-9484-270d82dc34e0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.343540] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Lock "c4883f84-e62b-4b59-9484-270d82dc34e0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.344385] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Acquiring lock "c4883f84-e62b-4b59-9484-270d82dc34e0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.344385] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Lock "c4883f84-e62b-4b59-9484-270d82dc34e0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.344385] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Lock "c4883f84-e62b-4b59-9484-270d82dc34e0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.348818] env[61857]: INFO nova.compute.manager [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Terminating instance [ 813.353437] env[61857]: DEBUG nova.compute.manager [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 813.353659] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.353908] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38e4088f-42c0-4e9e-923a-eb980b256d61 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.364166] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for the task: (returnval){ [ 813.364166] env[61857]: value = "task-950935" [ 813.364166] env[61857]: _type = "Task" [ 813.364166] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.371589] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950935, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.553291] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69dd92f0-8e76-4c19-bfab-2e564a83ac07 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.562155] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210a5342-a3bb-4bc1-ab98-517933456967 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.568466] env[61857]: DEBUG nova.network.neutron [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Updating instance_info_cache with network_info: [{"id": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "address": "fa:16:3e:88:8c:fd", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea2b08a8-d6", "ovs_interfaceid": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.598677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.598677] env[61857]: DEBUG nova.compute.manager [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Instance network_info: |[{"id": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "address": "fa:16:3e:88:8c:fd", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea2b08a8-d6", "ovs_interfaceid": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 813.599398] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cc9a68-241d-4c70-af54-b92b20f009b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.602582] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:8c:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ffea58a-e043-4ad1-9ad7-45a61fdca001', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea2b08a8-d692-400f-a7a5-124e3a5c14de', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.610019] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Creating folder: Project (5653ed6ab0a948ef89bb03cf7426dd10). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 813.610600] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-43ed81ff-b42c-44fe-bec1-378668c81ff8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.620083] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df514fd-a630-46d6-be9f-cca59129c0c9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.627722] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Created folder: Project (5653ed6ab0a948ef89bb03cf7426dd10) in parent group-v214027. [ 813.628009] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Creating folder: Instances. Parent ref: group-v214101. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 813.628804] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5d1c069-1d39-43b8-96fc-23075823abe6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.643675] env[61857]: DEBUG nova.compute.provider_tree [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 813.654601] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Created folder: Instances in parent group-v214101. [ 813.654976] env[61857]: DEBUG oslo.service.loopingcall [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.655283] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 813.655550] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0a09a4ef-8da8-403b-90e1-8c6cdc6998fb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.684332] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.684332] env[61857]: value = "task-950938" [ 813.684332] env[61857]: _type = "Task" [ 813.684332] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.694051] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950938, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.773281] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950934, 'name': PowerOffVM_Task, 'duration_secs': 0.260172} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.774076] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 813.781805] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Reconfiguring VM instance instance-00000036 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 813.782390] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ce67d8e-77a3-4655-97e8-2a71486a4a08 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.802979] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 813.802979] env[61857]: value = "task-950939" [ 813.802979] env[61857]: _type = "Task" [ 813.802979] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.813524] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950939, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.876181] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950935, 'name': PowerOffVM_Task, 'duration_secs': 0.183813} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.876769] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 813.877184] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Volume detach. Driver type: vmdk {{(pid=61857) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 813.877513] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214043', 'volume_id': '3583f165-a169-43a4-975e-46e6fb77fbef', 'name': 'volume-3583f165-a169-43a4-975e-46e6fb77fbef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4883f84-e62b-4b59-9484-270d82dc34e0', 'attached_at': '', 'detached_at': '', 'volume_id': '3583f165-a169-43a4-975e-46e6fb77fbef', 'serial': '3583f165-a169-43a4-975e-46e6fb77fbef'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 813.878569] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9238b61a-6c50-4bf6-b1cc-ed9eed2d4b1e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.901964] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18878c63-65fa-44b3-819a-ad3e33707344 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.910897] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746839a1-a0ba-4622-9e48-462520961ba6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.933932] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d830a50a-259f-45d0-a369-922972db5701 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.954056] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] The volume has not been displaced from its original location: [datastore2] volume-3583f165-a169-43a4-975e-46e6fb77fbef/volume-3583f165-a169-43a4-975e-46e6fb77fbef.vmdk. No consolidation needed. {{(pid=61857) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 813.961063] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Reconfiguring VM instance instance-00000033 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 813.963047] env[61857]: DEBUG nova.compute.manager [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.965312] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48f524b2-f19f-4d66-9c4f-a572227e1c76 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.990931] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for the task: (returnval){ [ 813.990931] env[61857]: value = "task-950940" [ 813.990931] env[61857]: _type = "Task" [ 813.990931] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.004634] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950940, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.006953] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.007187] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.007359] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.007555] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.007732] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.007891] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.008129] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.008337] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.008517] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.008710] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.008892] env[61857]: DEBUG nova.virt.hardware [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.009770] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74fe91e-71f5-47e4-822a-33b4c9122bcd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.017373] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bb7c7a-5906-4676-a6e5-2555f04df248 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.179991] env[61857]: DEBUG nova.scheduler.client.report [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Updated inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with generation 88 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 814.180426] env[61857]: DEBUG nova.compute.provider_tree [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Updating resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 generation from 88 to 89 during operation: update_inventory {{(pid=61857) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 814.180601] env[61857]: DEBUG nova.compute.provider_tree [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 814.194909] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950938, 'name': CreateVM_Task, 'duration_secs': 0.349773} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.196405] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 814.197898] env[61857]: DEBUG nova.compute.manager [req-f6a33964-a2de-4501-bec2-1a2834782e5f req-2107b64a-6b37-4c76-a0f8-093c7326e375 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Received event network-changed-ea2b08a8-d692-400f-a7a5-124e3a5c14de {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.198789] env[61857]: DEBUG nova.compute.manager [req-f6a33964-a2de-4501-bec2-1a2834782e5f req-2107b64a-6b37-4c76-a0f8-093c7326e375 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Refreshing instance network info cache due to event network-changed-ea2b08a8-d692-400f-a7a5-124e3a5c14de. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 814.198789] env[61857]: DEBUG oslo_concurrency.lockutils [req-f6a33964-a2de-4501-bec2-1a2834782e5f req-2107b64a-6b37-4c76-a0f8-093c7326e375 service nova] Acquiring lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.198789] env[61857]: DEBUG oslo_concurrency.lockutils [req-f6a33964-a2de-4501-bec2-1a2834782e5f req-2107b64a-6b37-4c76-a0f8-093c7326e375 service nova] Acquired lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.198789] env[61857]: DEBUG nova.network.neutron [req-f6a33964-a2de-4501-bec2-1a2834782e5f req-2107b64a-6b37-4c76-a0f8-093c7326e375 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Refreshing network info cache for port ea2b08a8-d692-400f-a7a5-124e3a5c14de {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 814.201310] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.201575] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.202109] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.202956] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-571807ec-0946-4aec-8a6f-d372007de998 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.209266] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 814.209266] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52f69909-75e6-1e27-7223-de923029ed3c" [ 814.209266] env[61857]: _type = "Task" [ 814.209266] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.220852] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f69909-75e6-1e27-7223-de923029ed3c, 'name': SearchDatastore_Task, 'duration_secs': 0.00946} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.221318] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.221574] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.221914] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.222172] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.222398] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.222762] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b008eae0-20b9-4f94-87b5-f0c100baa41b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.231415] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.231676] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 814.232428] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2bdd892-0b8d-4486-a495-ef550d026636 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.240265] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 814.240265] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]529c4574-ab8a-9cbb-16a3-4dc42027a1c1" [ 814.240265] env[61857]: _type = "Task" [ 814.240265] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.248446] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529c4574-ab8a-9cbb-16a3-4dc42027a1c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.313564] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950939, 'name': ReconfigVM_Task, 'duration_secs': 0.26249} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.314112] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Reconfigured VM instance instance-00000036 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 814.314112] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 814.314112] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6ebf5e3-8c97-4d50-9977-c3a6093d45c5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.321565] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 814.321565] env[61857]: value = "task-950941" [ 814.321565] env[61857]: _type = "Task" [ 814.321565] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.329673] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950941, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.503201] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950940, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.655845] env[61857]: DEBUG nova.network.neutron [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Successfully updated port: 7ae4b256-08e6-483a-a76a-1c1bcc44e67d {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.685765] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.754s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.686068] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 814.688853] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.936s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.689143] env[61857]: DEBUG nova.objects.instance [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lazy-loading 'resources' on Instance uuid 11adaee5-8e56-4679-8373-5d5690a44ca9 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.752276] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529c4574-ab8a-9cbb-16a3-4dc42027a1c1, 'name': SearchDatastore_Task, 'duration_secs': 0.0087} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.753087] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b814d484-e809-407d-ae0c-429f1f0ebb9f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.759024] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 814.759024] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52f02cb6-9982-b7fb-aa35-c995de91bb5a" [ 814.759024] env[61857]: _type = "Task" [ 814.759024] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.766951] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f02cb6-9982-b7fb-aa35-c995de91bb5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.833543] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950941, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.959973] env[61857]: DEBUG nova.network.neutron [req-f6a33964-a2de-4501-bec2-1a2834782e5f req-2107b64a-6b37-4c76-a0f8-093c7326e375 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Updated VIF entry in instance network info cache for port ea2b08a8-d692-400f-a7a5-124e3a5c14de. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 814.960615] env[61857]: DEBUG nova.network.neutron [req-f6a33964-a2de-4501-bec2-1a2834782e5f req-2107b64a-6b37-4c76-a0f8-093c7326e375 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Updating instance_info_cache with network_info: [{"id": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "address": "fa:16:3e:88:8c:fd", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea2b08a8-d6", "ovs_interfaceid": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.002375] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950940, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.159623] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "refresh_cache-93e03e35-15a0-49e7-b1be-09178eabbfda" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.159871] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "refresh_cache-93e03e35-15a0-49e7-b1be-09178eabbfda" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.160023] env[61857]: DEBUG nova.network.neutron [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 815.192791] env[61857]: DEBUG nova.compute.utils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.197791] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 815.197981] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 815.274526] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f02cb6-9982-b7fb-aa35-c995de91bb5a, 'name': SearchDatastore_Task, 'duration_secs': 0.009448} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.274526] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.274526] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 96ff157c-4471-4a12-ad99-0aafd6c2dce5/96ff157c-4471-4a12-ad99-0aafd6c2dce5.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 815.274526] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1555969-d385-4b16-a5f7-d533ecf4d2f1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.284127] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 815.284127] env[61857]: value = "task-950942" [ 815.284127] env[61857]: _type = "Task" [ 815.284127] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.293658] env[61857]: DEBUG nova.policy [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a3a9944299d4574bf46978852f11d92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b21cc6a46cc04f20b942603dc80ec6c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.301406] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-950942, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.332743] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950941, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.464875] env[61857]: DEBUG oslo_concurrency.lockutils [req-f6a33964-a2de-4501-bec2-1a2834782e5f req-2107b64a-6b37-4c76-a0f8-093c7326e375 service nova] Releasing lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.504671] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950940, 'name': ReconfigVM_Task, 'duration_secs': 1.255112} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.508077] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Reconfigured VM instance instance-00000033 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 815.515357] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71f04f88-b362-4757-a160-b3054f357e46 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.538045] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for the task: (returnval){ [ 815.538045] env[61857]: value = "task-950943" [ 815.538045] env[61857]: _type = "Task" [ 815.538045] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.551772] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950943, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.655766] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Successfully created port: 93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.700929] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.708586] env[61857]: DEBUG nova.network.neutron [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 815.733024] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad460d8-698d-4665-b275-d3982c479435 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.741552] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516b29ac-b97d-4d41-9411-d681270688ca {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.774973] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd5658a-b8f0-41f7-8079-30c0b2dd6f11 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.789186] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c322c140-b99d-4903-81dc-3aa14025f2df {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.798342] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-950942, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501628} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.806297] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 96ff157c-4471-4a12-ad99-0aafd6c2dce5/96ff157c-4471-4a12-ad99-0aafd6c2dce5.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 815.806547] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 815.807232] env[61857]: DEBUG nova.compute.provider_tree [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 815.808149] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7f54d99-ed0c-41cf-8d9a-778f9e382495 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.817728] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 815.817728] env[61857]: value = "task-950944" [ 815.817728] env[61857]: _type = "Task" [ 815.817728] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.830237] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-950944, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.836021] env[61857]: DEBUG oslo_vmware.api [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-950941, 'name': PowerOnVM_Task, 'duration_secs': 1.408907} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.836021] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 815.836021] env[61857]: DEBUG nova.compute.manager [None req-a3464290-8e54-48ba-8c4c-45e1b2af8718 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 815.836749] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44675e4e-9691-40fc-a1ee-24a2d41db90f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.999792] env[61857]: DEBUG nova.network.neutron [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Updating instance_info_cache with network_info: [{"id": "7ae4b256-08e6-483a-a76a-1c1bcc44e67d", "address": "fa:16:3e:3e:1a:c4", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ae4b256-08", "ovs_interfaceid": "7ae4b256-08e6-483a-a76a-1c1bcc44e67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.054310] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950943, 'name': ReconfigVM_Task, 'duration_secs': 0.185558} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.054310] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214043', 'volume_id': '3583f165-a169-43a4-975e-46e6fb77fbef', 'name': 'volume-3583f165-a169-43a4-975e-46e6fb77fbef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4883f84-e62b-4b59-9484-270d82dc34e0', 'attached_at': '', 'detached_at': '', 'volume_id': '3583f165-a169-43a4-975e-46e6fb77fbef', 'serial': '3583f165-a169-43a4-975e-46e6fb77fbef'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 816.054688] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 816.055888] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda97d79-a8b0-4f26-9086-4d965392652e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.065678] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 816.065818] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-73530a19-55d6-4ac6-a139-4e8904a940c3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.135905] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 816.136163] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 816.136352] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Deleting the datastore file [datastore2] c4883f84-e62b-4b59-9484-270d82dc34e0 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.136742] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-43a35a71-5519-4359-b51b-22061317811b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.144115] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for the task: (returnval){ [ 816.144115] env[61857]: value = "task-950946" [ 816.144115] env[61857]: _type = "Task" [ 816.144115] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.152639] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.224421] env[61857]: DEBUG nova.compute.manager [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Received event network-vif-plugged-7ae4b256-08e6-483a-a76a-1c1bcc44e67d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 816.224421] env[61857]: DEBUG oslo_concurrency.lockutils [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] Acquiring lock "93e03e35-15a0-49e7-b1be-09178eabbfda-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.224576] env[61857]: DEBUG oslo_concurrency.lockutils [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] Lock "93e03e35-15a0-49e7-b1be-09178eabbfda-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.224644] env[61857]: DEBUG oslo_concurrency.lockutils [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] Lock "93e03e35-15a0-49e7-b1be-09178eabbfda-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.224938] env[61857]: DEBUG nova.compute.manager [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] No waiting events found dispatching network-vif-plugged-7ae4b256-08e6-483a-a76a-1c1bcc44e67d {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 816.225017] env[61857]: WARNING nova.compute.manager [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Received unexpected event network-vif-plugged-7ae4b256-08e6-483a-a76a-1c1bcc44e67d for instance with vm_state building and task_state spawning. [ 816.225341] env[61857]: DEBUG nova.compute.manager [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Received event network-changed-7ae4b256-08e6-483a-a76a-1c1bcc44e67d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 816.225341] env[61857]: DEBUG nova.compute.manager [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Refreshing instance network info cache due to event network-changed-7ae4b256-08e6-483a-a76a-1c1bcc44e67d. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 816.225531] env[61857]: DEBUG oslo_concurrency.lockutils [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] Acquiring lock "refresh_cache-93e03e35-15a0-49e7-b1be-09178eabbfda" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.327029] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-950944, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.146601} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.327318] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 816.328103] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead4e1d8-5288-4615-90f3-1a603748dcb9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.350066] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 96ff157c-4471-4a12-ad99-0aafd6c2dce5/96ff157c-4471-4a12-ad99-0aafd6c2dce5.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.351170] env[61857]: DEBUG nova.scheduler.client.report [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Updated inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with generation 89 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 816.351457] env[61857]: DEBUG nova.compute.provider_tree [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Updating resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 generation from 89 to 90 during operation: update_inventory {{(pid=61857) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 816.351658] env[61857]: DEBUG nova.compute.provider_tree [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 816.356758] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eec5a19b-fa4a-4b3f-ae42-003c57bd6b2b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.381539] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 816.381539] env[61857]: value = "task-950947" [ 816.381539] env[61857]: _type = "Task" [ 816.381539] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.391219] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-950947, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.502576] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "refresh_cache-93e03e35-15a0-49e7-b1be-09178eabbfda" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.503000] env[61857]: DEBUG nova.compute.manager [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Instance network_info: |[{"id": "7ae4b256-08e6-483a-a76a-1c1bcc44e67d", "address": "fa:16:3e:3e:1a:c4", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ae4b256-08", "ovs_interfaceid": "7ae4b256-08e6-483a-a76a-1c1bcc44e67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.503408] env[61857]: DEBUG oslo_concurrency.lockutils [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] Acquired lock "refresh_cache-93e03e35-15a0-49e7-b1be-09178eabbfda" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.503673] env[61857]: DEBUG nova.network.neutron [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Refreshing network info cache for port 7ae4b256-08e6-483a-a76a-1c1bcc44e67d {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 816.505332] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:1a:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7ae4b256-08e6-483a-a76a-1c1bcc44e67d', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.517737] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Creating folder: Project (62bd314daba846cd9ed3dc994dd1034b). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.519679] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6cd400a6-d457-40f3-84db-19f23fe4b717 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.532528] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Created folder: Project (62bd314daba846cd9ed3dc994dd1034b) in parent group-v214027. [ 816.532809] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Creating folder: Instances. Parent ref: group-v214104. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.533139] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9db6d977-6b74-42b7-8f99-ba1a7536fd88 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.542958] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Created folder: Instances in parent group-v214104. [ 816.543297] env[61857]: DEBUG oslo.service.loopingcall [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.543440] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 816.543732] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53800d2e-c1b5-4c36-982a-46a8b4cb4614 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.563130] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.563130] env[61857]: value = "task-950950" [ 816.563130] env[61857]: _type = "Task" [ 816.563130] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.571471] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950950, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.655953] env[61857]: DEBUG oslo_vmware.api [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Task: {'id': task-950946, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108946} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.656332] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.656562] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 816.656794] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 816.657028] env[61857]: INFO nova.compute.manager [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Took 3.30 seconds to destroy the instance on the hypervisor. [ 816.657348] env[61857]: DEBUG oslo.service.loopingcall [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.657562] env[61857]: DEBUG nova.compute.manager [-] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 816.657703] env[61857]: DEBUG nova.network.neutron [-] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 816.709709] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.738747] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.739051] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.739468] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.739468] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.739633] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.739811] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.740076] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.740300] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.740488] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.740691] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.740900] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.741888] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc50b92c-361c-44c3-9993-413cc989bc7e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.750336] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c85bfb7-9353-499c-afe2-2e8921309a2c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.875531] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.184s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.876514] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.277s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.877990] env[61857]: INFO nova.compute.claims [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.891495] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-950947, 'name': ReconfigVM_Task, 'duration_secs': 0.469463} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.891853] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 96ff157c-4471-4a12-ad99-0aafd6c2dce5/96ff157c-4471-4a12-ad99-0aafd6c2dce5.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.892429] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc2ba6e1-1756-455f-8051-10208d92d432 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.900940] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 816.900940] env[61857]: value = "task-950951" [ 816.900940] env[61857]: _type = "Task" [ 816.900940] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.910864] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-950951, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.912037] env[61857]: INFO nova.scheduler.client.report [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Deleted allocations for instance 11adaee5-8e56-4679-8373-5d5690a44ca9 [ 817.075564] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950950, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.418048] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-950951, 'name': Rename_Task, 'duration_secs': 0.152} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.420700] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 817.423632] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0e6e653c-9f03-4369-9e9e-873bafa2fea5 tempest-ServersTestJSON-2111788929 tempest-ServersTestJSON-2111788929-project-member] Lock "11adaee5-8e56-4679-8373-5d5690a44ca9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.154s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.424534] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b89cdbe-94f1-4e02-8b75-67b616d21f97 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.432456] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Successfully updated port: 93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.440102] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 817.440102] env[61857]: value = "task-950952" [ 817.440102] env[61857]: _type = "Task" [ 817.440102] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.458113] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-950952, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.514256] env[61857]: DEBUG nova.network.neutron [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Updated VIF entry in instance network info cache for port 7ae4b256-08e6-483a-a76a-1c1bcc44e67d. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 817.514757] env[61857]: DEBUG nova.network.neutron [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Updating instance_info_cache with network_info: [{"id": "7ae4b256-08e6-483a-a76a-1c1bcc44e67d", "address": "fa:16:3e:3e:1a:c4", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ae4b256-08", "ovs_interfaceid": "7ae4b256-08e6-483a-a76a-1c1bcc44e67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.522346] env[61857]: DEBUG nova.compute.manager [req-9d3b688d-e803-4b57-a098-6be26b504e07 req-26d92a7c-5d54-4031-ba9c-bee2aa913073 service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Received event network-vif-deleted-e9861f7d-8566-414c-a73c-1e3caf722cd2 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 817.522661] env[61857]: INFO nova.compute.manager [req-9d3b688d-e803-4b57-a098-6be26b504e07 req-26d92a7c-5d54-4031-ba9c-bee2aa913073 service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Neutron deleted interface e9861f7d-8566-414c-a73c-1e3caf722cd2; detaching it from the instance and deleting it from the info cache [ 817.522909] env[61857]: DEBUG nova.network.neutron [req-9d3b688d-e803-4b57-a098-6be26b504e07 req-26d92a7c-5d54-4031-ba9c-bee2aa913073 service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.578207] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950950, 'name': CreateVM_Task, 'duration_secs': 0.718349} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.579777] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 817.579777] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.579777] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.579957] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.580366] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a69a8747-08d9-4c01-b82b-a655500ca75f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.585587] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 817.585587] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52eb7c8d-1ec7-0f37-1a86-6fa74cf190ba" [ 817.585587] env[61857]: _type = "Task" [ 817.585587] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.599812] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52eb7c8d-1ec7-0f37-1a86-6fa74cf190ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.934639] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "refresh_cache-bf2f46d4-5c6e-4493-b087-fe5a316b8c88" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.935355] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired lock "refresh_cache-bf2f46d4-5c6e-4493-b087-fe5a316b8c88" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.935355] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 817.950784] env[61857]: DEBUG nova.network.neutron [-] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.963088] env[61857]: DEBUG oslo_vmware.api [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-950952, 'name': PowerOnVM_Task, 'duration_secs': 0.478915} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.965705] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 817.965962] env[61857]: INFO nova.compute.manager [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Took 6.63 seconds to spawn the instance on the hypervisor. [ 817.966406] env[61857]: DEBUG nova.compute.manager [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.968307] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67365f4d-a7b8-42af-aa91-0ffaf2ce78dd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.018699] env[61857]: DEBUG oslo_concurrency.lockutils [req-7c4b37a1-c028-4712-ae19-df0008b02539 req-694cca33-97d1-4b72-9af9-41017bdeeeda service nova] Releasing lock "refresh_cache-93e03e35-15a0-49e7-b1be-09178eabbfda" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.026786] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34bf3f61-8da0-422c-a788-99aa0d1d0261 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.038681] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb6463f-49ee-47e7-9db7-537a21d484a5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.083338] env[61857]: DEBUG nova.compute.manager [req-9d3b688d-e803-4b57-a098-6be26b504e07 req-26d92a7c-5d54-4031-ba9c-bee2aa913073 service nova] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Detach interface failed, port_id=e9861f7d-8566-414c-a73c-1e3caf722cd2, reason: Instance c4883f84-e62b-4b59-9484-270d82dc34e0 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 818.098215] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52eb7c8d-1ec7-0f37-1a86-6fa74cf190ba, 'name': SearchDatastore_Task, 'duration_secs': 0.016054} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.098627] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.098902] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 818.099196] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.099406] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.099659] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 818.102020] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ea5190e-d9f1-4e56-ba32-2e8c5e1b1d8a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.109718] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 818.109960] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 818.111354] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d63c0828-86fa-4cc9-a163-2557b7165e42 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.117705] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 818.117705] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523d27d2-cca6-fd2a-3d58-1760980a279b" [ 818.117705] env[61857]: _type = "Task" [ 818.117705] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.129353] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523d27d2-cca6-fd2a-3d58-1760980a279b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.298958] env[61857]: DEBUG nova.compute.manager [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Received event network-vif-plugged-93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.298958] env[61857]: DEBUG oslo_concurrency.lockutils [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] Acquiring lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.298958] env[61857]: DEBUG oslo_concurrency.lockutils [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] Lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.299392] env[61857]: DEBUG oslo_concurrency.lockutils [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] Lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.299392] env[61857]: DEBUG nova.compute.manager [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] No waiting events found dispatching network-vif-plugged-93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 818.299475] env[61857]: WARNING nova.compute.manager [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Received unexpected event network-vif-plugged-93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb for instance with vm_state building and task_state spawning. [ 818.299817] env[61857]: DEBUG nova.compute.manager [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Received event network-changed-93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.300026] env[61857]: DEBUG nova.compute.manager [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Refreshing instance network info cache due to event network-changed-93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 818.300228] env[61857]: DEBUG oslo_concurrency.lockutils [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] Acquiring lock "refresh_cache-bf2f46d4-5c6e-4493-b087-fe5a316b8c88" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.411152] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa2cf41f-a7ad-47de-a2a1-e4979d396d59 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.419815] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30b43a4-356a-4cc1-8fca-a95853ce96f0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.456420] env[61857]: INFO nova.compute.manager [-] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Took 1.80 seconds to deallocate network for instance. [ 818.457798] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7ebd57-4204-45f8-9dbf-e20f0ef47b40 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.468769] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5f441c-4060-4d0e-9daf-b144dba3af51 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.492598] env[61857]: DEBUG nova.compute.provider_tree [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.499511] env[61857]: INFO nova.compute.manager [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Took 41.10 seconds to build instance. [ 818.505553] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 818.626990] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523d27d2-cca6-fd2a-3d58-1760980a279b, 'name': SearchDatastore_Task, 'duration_secs': 0.012499} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.627784] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d818c0e-4023-47cb-a668-df84ccd08b18 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.633328] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 818.633328] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]526aeafa-af94-b88c-a319-f287879d2e36" [ 818.633328] env[61857]: _type = "Task" [ 818.633328] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.643865] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]526aeafa-af94-b88c-a319-f287879d2e36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.777574] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Updating instance_info_cache with network_info: [{"id": "93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb", "address": "fa:16:3e:c8:fe:f8", "network": {"id": "a2079895-c2b5-4cb6-9429-3990ae7b910e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-293337110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b21cc6a46cc04f20b942603dc80ec6c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93a6a35f-96", "ovs_interfaceid": "93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.001668] env[61857]: DEBUG oslo_concurrency.lockutils [None req-09358970-e9f8-4317-ba36-2a4bd85e50fe tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.022s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.003946] env[61857]: DEBUG nova.scheduler.client.report [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.021846] env[61857]: INFO nova.compute.manager [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Took 0.56 seconds to detach 1 volumes for instance. [ 819.023608] env[61857]: DEBUG nova.compute.manager [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Deleting volume: 3583f165-a169-43a4-975e-46e6fb77fbef {{(pid=61857) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 819.145043] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]526aeafa-af94-b88c-a319-f287879d2e36, 'name': SearchDatastore_Task, 'duration_secs': 0.019455} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.145393] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.145809] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 93e03e35-15a0-49e7-b1be-09178eabbfda/93e03e35-15a0-49e7-b1be-09178eabbfda.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 819.146110] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c272b944-00cd-4198-be75-8bab96989dd7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.154803] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 819.154803] env[61857]: value = "task-950954" [ 819.154803] env[61857]: _type = "Task" [ 819.154803] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.165404] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950954, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.280908] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Releasing lock "refresh_cache-bf2f46d4-5c6e-4493-b087-fe5a316b8c88" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.281493] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Instance network_info: |[{"id": "93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb", "address": "fa:16:3e:c8:fe:f8", "network": {"id": "a2079895-c2b5-4cb6-9429-3990ae7b910e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-293337110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b21cc6a46cc04f20b942603dc80ec6c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93a6a35f-96", "ovs_interfaceid": "93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 819.281874] env[61857]: DEBUG oslo_concurrency.lockutils [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] Acquired lock "refresh_cache-bf2f46d4-5c6e-4493-b087-fe5a316b8c88" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.282131] env[61857]: DEBUG nova.network.neutron [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Refreshing network info cache for port 93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.283390] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:fe:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69744f59-ecac-4b0b-831e-82a274d7acbb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 819.294176] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Creating folder: Project (b21cc6a46cc04f20b942603dc80ec6c4). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 819.295011] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-48f88f4d-ca5f-41ea-8e9e-8d72edead4c2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.305059] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Created folder: Project (b21cc6a46cc04f20b942603dc80ec6c4) in parent group-v214027. [ 819.305280] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Creating folder: Instances. Parent ref: group-v214107. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 819.305531] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be4ccede-9b40-4056-ae97-180fb6237482 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.323830] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Created folder: Instances in parent group-v214107. [ 819.324116] env[61857]: DEBUG oslo.service.loopingcall [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.324434] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 819.324539] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dbfb2f10-adcd-4562-9f4b-e41d823706dd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.354190] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 819.354190] env[61857]: value = "task-950957" [ 819.354190] env[61857]: _type = "Task" [ 819.354190] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.362788] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950957, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.516196] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.639s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.516836] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.520391] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 819.524992] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.796s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.527886] env[61857]: INFO nova.compute.claims [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.579471] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.669960] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950954, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.870834] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950957, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.026408] env[61857]: DEBUG nova.compute.utils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.028154] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.028386] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 820.064434] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.153532] env[61857]: DEBUG nova.policy [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a3a9944299d4574bf46978852f11d92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b21cc6a46cc04f20b942603dc80ec6c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.169279] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950954, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582673} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.169679] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 93e03e35-15a0-49e7-b1be-09178eabbfda/93e03e35-15a0-49e7-b1be-09178eabbfda.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 820.169998] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.170361] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df97b026-7f47-40ec-8188-271bdb3e3bae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.177154] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 820.177154] env[61857]: value = "task-950958" [ 820.177154] env[61857]: _type = "Task" [ 820.177154] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.187439] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950958, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.340224] env[61857]: DEBUG nova.network.neutron [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Updated VIF entry in instance network info cache for port 93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 820.340224] env[61857]: DEBUG nova.network.neutron [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Updating instance_info_cache with network_info: [{"id": "93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb", "address": "fa:16:3e:c8:fe:f8", "network": {"id": "a2079895-c2b5-4cb6-9429-3990ae7b910e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-293337110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b21cc6a46cc04f20b942603dc80ec6c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93a6a35f-96", "ovs_interfaceid": "93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.369720] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950957, 'name': CreateVM_Task, 'duration_secs': 0.591417} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.369931] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 820.370691] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.371081] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.371205] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 820.371871] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74698fa0-62c6-4db2-892a-91372ce3a043 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.377516] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 820.377516] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5231b396-d0e5-1e2f-2470-fa76955b4571" [ 820.377516] env[61857]: _type = "Task" [ 820.377516] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.386562] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5231b396-d0e5-1e2f-2470-fa76955b4571, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.531380] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.597662] env[61857]: DEBUG nova.compute.manager [req-9997070d-c710-418a-a283-0ec553fbf2d6 req-2f6d22f7-71a6-4a93-a9af-9cbcbe2c317f service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Received event network-changed-ea2b08a8-d692-400f-a7a5-124e3a5c14de {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.598024] env[61857]: DEBUG nova.compute.manager [req-9997070d-c710-418a-a283-0ec553fbf2d6 req-2f6d22f7-71a6-4a93-a9af-9cbcbe2c317f service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Refreshing instance network info cache due to event network-changed-ea2b08a8-d692-400f-a7a5-124e3a5c14de. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 820.599513] env[61857]: DEBUG oslo_concurrency.lockutils [req-9997070d-c710-418a-a283-0ec553fbf2d6 req-2f6d22f7-71a6-4a93-a9af-9cbcbe2c317f service nova] Acquiring lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.599513] env[61857]: DEBUG oslo_concurrency.lockutils [req-9997070d-c710-418a-a283-0ec553fbf2d6 req-2f6d22f7-71a6-4a93-a9af-9cbcbe2c317f service nova] Acquired lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.599513] env[61857]: DEBUG nova.network.neutron [req-9997070d-c710-418a-a283-0ec553fbf2d6 req-2f6d22f7-71a6-4a93-a9af-9cbcbe2c317f service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Refreshing network info cache for port ea2b08a8-d692-400f-a7a5-124e3a5c14de {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 820.693232] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950958, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069641} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.694146] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.695141] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cdfedf0-fc50-4dfa-9aa6-44292581865b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.728021] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 93e03e35-15a0-49e7-b1be-09178eabbfda/93e03e35-15a0-49e7-b1be-09178eabbfda.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.731179] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd3b0229-019a-4a0b-90ef-4a9f15620ed3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.757127] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 820.757127] env[61857]: value = "task-950959" [ 820.757127] env[61857]: _type = "Task" [ 820.757127] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.770541] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950959, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.771256] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Successfully created port: 3adcb286-015b-4b35-87f4-befacb5b4f3d {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.843013] env[61857]: DEBUG oslo_concurrency.lockutils [req-23ea90dc-5930-433f-ba62-7d12a890ea97 req-8608aac9-2bf8-4d77-adaa-f5103f9395a4 service nova] Releasing lock "refresh_cache-bf2f46d4-5c6e-4493-b087-fe5a316b8c88" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.891301] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5231b396-d0e5-1e2f-2470-fa76955b4571, 'name': SearchDatastore_Task, 'duration_secs': 0.029827} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.891635] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.891879] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.892138] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.892293] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.892469] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.892738] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85d20a94-abef-47a7-800c-f46ff458cec3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.905026] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.905251] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 820.906091] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c027a7ad-50e4-4458-9685-d7030e5627f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.915170] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 820.915170] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52dcb4d0-5df0-7f8f-0a68-2767a287a980" [ 820.915170] env[61857]: _type = "Task" [ 820.915170] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.926107] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52dcb4d0-5df0-7f8f-0a68-2767a287a980, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.945467] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.945779] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.177470] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35794bb2-9acb-4628-b7cf-b4b7da773123 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.188734] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39c8de7-f4a7-4b62-a988-2f4468942ba1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.230100] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e64b78-1411-4433-90a0-8ac6d0a483b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.238499] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca15d8b-5d48-490c-8a74-50521430d6ea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.258647] env[61857]: DEBUG nova.compute.provider_tree [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.269669] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950959, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.426763] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52dcb4d0-5df0-7f8f-0a68-2767a287a980, 'name': SearchDatastore_Task, 'duration_secs': 0.022889} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.430396] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd7ed0f9-27c5-4097-b514-9e472ca52f9d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.437684] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 821.437684] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]526379b2-e53b-97d4-7dbd-49e2c5455de2" [ 821.437684] env[61857]: _type = "Task" [ 821.437684] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.446460] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]526379b2-e53b-97d4-7dbd-49e2c5455de2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.481565] env[61857]: DEBUG nova.network.neutron [req-9997070d-c710-418a-a283-0ec553fbf2d6 req-2f6d22f7-71a6-4a93-a9af-9cbcbe2c317f service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Updated VIF entry in instance network info cache for port ea2b08a8-d692-400f-a7a5-124e3a5c14de. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 821.482209] env[61857]: DEBUG nova.network.neutron [req-9997070d-c710-418a-a283-0ec553fbf2d6 req-2f6d22f7-71a6-4a93-a9af-9cbcbe2c317f service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Updating instance_info_cache with network_info: [{"id": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "address": "fa:16:3e:88:8c:fd", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea2b08a8-d6", "ovs_interfaceid": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.543917] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.566881] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.567161] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.567333] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.567523] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.567677] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.567833] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.568085] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.568546] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.568802] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.569013] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.569212] env[61857]: DEBUG nova.virt.hardware [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.570117] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9aabde7-9983-4ba5-958b-06ea3c9e9c9c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.577983] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6122ae04-64d6-4d01-871e-82057671ec1c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.764865] env[61857]: DEBUG nova.scheduler.client.report [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.775401] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950959, 'name': ReconfigVM_Task, 'duration_secs': 0.7448} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.775401] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 93e03e35-15a0-49e7-b1be-09178eabbfda/93e03e35-15a0-49e7-b1be-09178eabbfda.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.775695] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe14bfd6-92b7-4c66-8af9-20ac82fd8ff8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.782926] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 821.782926] env[61857]: value = "task-950960" [ 821.782926] env[61857]: _type = "Task" [ 821.782926] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.793089] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950960, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.949457] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]526379b2-e53b-97d4-7dbd-49e2c5455de2, 'name': SearchDatastore_Task, 'duration_secs': 0.023924} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.949789] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.950104] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] bf2f46d4-5c6e-4493-b087-fe5a316b8c88/bf2f46d4-5c6e-4493-b087-fe5a316b8c88.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 821.950426] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cdba07a8-7f96-4a6b-abc4-2315cf0828b2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.957434] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 821.957434] env[61857]: value = "task-950961" [ 821.957434] env[61857]: _type = "Task" [ 821.957434] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.969695] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950961, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.985530] env[61857]: DEBUG oslo_concurrency.lockutils [req-9997070d-c710-418a-a283-0ec553fbf2d6 req-2f6d22f7-71a6-4a93-a9af-9cbcbe2c317f service nova] Releasing lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.271271] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.746s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.271620] env[61857]: DEBUG nova.compute.manager [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.275367] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.449s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.278499] env[61857]: INFO nova.compute.claims [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.293688] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950960, 'name': Rename_Task, 'duration_secs': 0.250494} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.293988] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.294359] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3eddeb4d-65e2-4f92-9e05-492af0cb8820 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.301513] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 822.301513] env[61857]: value = "task-950962" [ 822.301513] env[61857]: _type = "Task" [ 822.301513] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.310835] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950962, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.468864] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950961, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.624137] env[61857]: DEBUG nova.compute.manager [req-562fb866-d0dd-4caa-b13f-3c5fecf6db1f req-127587e7-d8af-467d-a604-ef07449beefe service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Received event network-vif-plugged-3adcb286-015b-4b35-87f4-befacb5b4f3d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.624137] env[61857]: DEBUG oslo_concurrency.lockutils [req-562fb866-d0dd-4caa-b13f-3c5fecf6db1f req-127587e7-d8af-467d-a604-ef07449beefe service nova] Acquiring lock "a73f3163-b6fc-4764-b76a-451f823b5808-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.624137] env[61857]: DEBUG oslo_concurrency.lockutils [req-562fb866-d0dd-4caa-b13f-3c5fecf6db1f req-127587e7-d8af-467d-a604-ef07449beefe service nova] Lock "a73f3163-b6fc-4764-b76a-451f823b5808-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.624906] env[61857]: DEBUG oslo_concurrency.lockutils [req-562fb866-d0dd-4caa-b13f-3c5fecf6db1f req-127587e7-d8af-467d-a604-ef07449beefe service nova] Lock "a73f3163-b6fc-4764-b76a-451f823b5808-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.625312] env[61857]: DEBUG nova.compute.manager [req-562fb866-d0dd-4caa-b13f-3c5fecf6db1f req-127587e7-d8af-467d-a604-ef07449beefe service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] No waiting events found dispatching network-vif-plugged-3adcb286-015b-4b35-87f4-befacb5b4f3d {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.625736] env[61857]: WARNING nova.compute.manager [req-562fb866-d0dd-4caa-b13f-3c5fecf6db1f req-127587e7-d8af-467d-a604-ef07449beefe service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Received unexpected event network-vif-plugged-3adcb286-015b-4b35-87f4-befacb5b4f3d for instance with vm_state building and task_state spawning. [ 822.717783] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Successfully updated port: 3adcb286-015b-4b35-87f4-befacb5b4f3d {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.782191] env[61857]: DEBUG nova.compute.utils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.785604] env[61857]: DEBUG nova.compute.manager [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.785861] env[61857]: DEBUG nova.network.neutron [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 822.814903] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950962, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.841958] env[61857]: DEBUG nova.policy [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1496582f8474b14baf63b215a80e64e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3688a3ca85044527a84ad1a7b201fbdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.969377] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950961, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582748} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.969592] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] bf2f46d4-5c6e-4493-b087-fe5a316b8c88/bf2f46d4-5c6e-4493-b087-fe5a316b8c88.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 822.969822] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.970088] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44880a42-7487-4f92-a9eb-c915ced02e4e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.975960] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 822.975960] env[61857]: value = "task-950963" [ 822.975960] env[61857]: _type = "Task" [ 822.975960] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.984374] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950963, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.175124] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "71920a62-090d-4df1-937e-87df3b043e28" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.175578] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.175804] env[61857]: DEBUG nova.compute.manager [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.176767] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76384a03-4467-49af-afc3-146efcdb5e66 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.183558] env[61857]: DEBUG nova.compute.manager [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61857) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 823.183918] env[61857]: DEBUG nova.objects.instance [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'flavor' on Instance uuid 71920a62-090d-4df1-937e-87df3b043e28 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 823.226334] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "refresh_cache-a73f3163-b6fc-4764-b76a-451f823b5808" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.226501] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired lock "refresh_cache-a73f3163-b6fc-4764-b76a-451f823b5808" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.226705] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 823.311311] env[61857]: DEBUG nova.compute.manager [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 823.326554] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950962, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.460616] env[61857]: DEBUG nova.network.neutron [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Successfully created port: 0d52a5bd-50d8-41e1-86aa-b1eec00d8702 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.492274] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950963, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132963} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.495957] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.495957] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c736c5-3f81-435a-92c5-36d09f9b44ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.527201] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] bf2f46d4-5c6e-4493-b087-fe5a316b8c88/bf2f46d4-5c6e-4493-b087-fe5a316b8c88.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.529274] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-495a8914-fc23-4b30-ac23-d1f04ec9ee8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.552187] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 823.552187] env[61857]: value = "task-950964" [ 823.552187] env[61857]: _type = "Task" [ 823.552187] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.563208] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950964, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.688608] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 823.688885] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10dbad28-3d97-4ca6-90eb-84ad093cfa73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.698414] env[61857]: DEBUG oslo_vmware.api [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 823.698414] env[61857]: value = "task-950965" [ 823.698414] env[61857]: _type = "Task" [ 823.698414] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.707627] env[61857]: DEBUG oslo_vmware.api [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950965, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.802035] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 823.815313] env[61857]: DEBUG oslo_vmware.api [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950962, 'name': PowerOnVM_Task, 'duration_secs': 1.336942} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.815598] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.815807] env[61857]: INFO nova.compute.manager [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Took 9.85 seconds to spawn the instance on the hypervisor. [ 823.816494] env[61857]: DEBUG nova.compute.manager [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.816832] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f21ee2-1497-4c85-8315-175968265ff4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.839446] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5225e14-9492-4354-981a-4244febb565b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.850113] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef892173-d3e9-4b21-adbd-b783bf9d1a89 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.897514] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84bd5ca-0884-4a71-a957-d001d7fd6655 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.906924] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe7ae54-b6cd-4d12-8327-5fce520b1a8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.922826] env[61857]: DEBUG nova.compute.provider_tree [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.051023] env[61857]: DEBUG nova.network.neutron [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Updating instance_info_cache with network_info: [{"id": "3adcb286-015b-4b35-87f4-befacb5b4f3d", "address": "fa:16:3e:0f:4f:d3", "network": {"id": "a2079895-c2b5-4cb6-9429-3990ae7b910e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-293337110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b21cc6a46cc04f20b942603dc80ec6c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3adcb286-01", "ovs_interfaceid": "3adcb286-015b-4b35-87f4-befacb5b4f3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.060620] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950964, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.176455] env[61857]: DEBUG oslo_vmware.rw_handles [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52737eee-4b9a-07b9-31bc-65759d471c5c/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 824.177377] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3220521d-0c2b-464a-a443-3e1f5d6bd833 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.183994] env[61857]: DEBUG oslo_vmware.rw_handles [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52737eee-4b9a-07b9-31bc-65759d471c5c/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 824.184203] env[61857]: ERROR oslo_vmware.rw_handles [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52737eee-4b9a-07b9-31bc-65759d471c5c/disk-0.vmdk due to incomplete transfer. [ 824.184437] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-786c392f-044e-4ed3-9533-956239ad6f15 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.190986] env[61857]: DEBUG oslo_vmware.rw_handles [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52737eee-4b9a-07b9-31bc-65759d471c5c/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 824.191205] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Uploaded image 3ab5d30a-b778-412e-a710-7edebe17a33f to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 824.193495] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 824.194015] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-23916110-7325-4cee-a34e-437135c263da {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.199409] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 824.199409] env[61857]: value = "task-950966" [ 824.199409] env[61857]: _type = "Task" [ 824.199409] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.209846] env[61857]: DEBUG oslo_vmware.api [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950965, 'name': PowerOffVM_Task, 'duration_secs': 0.261563} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.212794] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 824.212985] env[61857]: DEBUG nova.compute.manager [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.213280] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950966, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.214185] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034b040c-510f-450c-9e93-6ed62e0784a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.325330] env[61857]: DEBUG nova.compute.manager [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.347258] env[61857]: INFO nova.compute.manager [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Took 42.74 seconds to build instance. [ 824.351170] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.351170] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.351170] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.351170] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.351377] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.351524] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.351746] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.352412] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.352412] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.352412] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.352676] env[61857]: DEBUG nova.virt.hardware [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.353340] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d638676c-4f1c-4354-b7b7-f7f10292f4c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.362171] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-039cb44d-9f3c-49b6-b8ca-cdf05550ea80 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.425861] env[61857]: DEBUG nova.scheduler.client.report [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.555387] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Releasing lock "refresh_cache-a73f3163-b6fc-4764-b76a-451f823b5808" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.555716] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Instance network_info: |[{"id": "3adcb286-015b-4b35-87f4-befacb5b4f3d", "address": "fa:16:3e:0f:4f:d3", "network": {"id": "a2079895-c2b5-4cb6-9429-3990ae7b910e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-293337110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b21cc6a46cc04f20b942603dc80ec6c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3adcb286-01", "ovs_interfaceid": "3adcb286-015b-4b35-87f4-befacb5b4f3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 824.556176] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:4f:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69744f59-ecac-4b0b-831e-82a274d7acbb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3adcb286-015b-4b35-87f4-befacb5b4f3d', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.563753] env[61857]: DEBUG oslo.service.loopingcall [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.564534] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 824.564783] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98557fae-6c20-46b3-8d94-0e34fa9cf208 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.584293] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950964, 'name': ReconfigVM_Task, 'duration_secs': 0.78567} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.585014] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Reconfigured VM instance instance-0000003f to attach disk [datastore1] bf2f46d4-5c6e-4493-b087-fe5a316b8c88/bf2f46d4-5c6e-4493-b087-fe5a316b8c88.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.585760] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ddf9199-66df-44c6-947f-4c04f1b4009b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.588907] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.588907] env[61857]: value = "task-950967" [ 824.588907] env[61857]: _type = "Task" [ 824.588907] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.593029] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 824.593029] env[61857]: value = "task-950968" [ 824.593029] env[61857]: _type = "Task" [ 824.593029] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.599144] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950967, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.604635] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950968, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.679287] env[61857]: DEBUG nova.compute.manager [req-d1509e6b-f81d-4b47-84a6-116aac791f03 req-2f04f7c0-07fc-450f-a25d-1cb9b5d1bfa9 service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Received event network-changed-3adcb286-015b-4b35-87f4-befacb5b4f3d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.679585] env[61857]: DEBUG nova.compute.manager [req-d1509e6b-f81d-4b47-84a6-116aac791f03 req-2f04f7c0-07fc-450f-a25d-1cb9b5d1bfa9 service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Refreshing instance network info cache due to event network-changed-3adcb286-015b-4b35-87f4-befacb5b4f3d. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 824.679788] env[61857]: DEBUG oslo_concurrency.lockutils [req-d1509e6b-f81d-4b47-84a6-116aac791f03 req-2f04f7c0-07fc-450f-a25d-1cb9b5d1bfa9 service nova] Acquiring lock "refresh_cache-a73f3163-b6fc-4764-b76a-451f823b5808" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.679899] env[61857]: DEBUG oslo_concurrency.lockutils [req-d1509e6b-f81d-4b47-84a6-116aac791f03 req-2f04f7c0-07fc-450f-a25d-1cb9b5d1bfa9 service nova] Acquired lock "refresh_cache-a73f3163-b6fc-4764-b76a-451f823b5808" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.680121] env[61857]: DEBUG nova.network.neutron [req-d1509e6b-f81d-4b47-84a6-116aac791f03 req-2f04f7c0-07fc-450f-a25d-1cb9b5d1bfa9 service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Refreshing network info cache for port 3adcb286-015b-4b35-87f4-befacb5b4f3d {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 824.714876] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950966, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.724927] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6fd27978-23d5-4857-9c77-ba755f98faf3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.849850] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6b9c7c3c-0ada-471d-b08f-f2d31703f39b tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "93e03e35-15a0-49e7-b1be-09178eabbfda" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.859s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.933025] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.658s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.933541] env[61857]: DEBUG nova.compute.manager [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 824.938713] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.746s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.942179] env[61857]: INFO nova.compute.claims [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.102695] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950967, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.105822] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950968, 'name': Rename_Task, 'duration_secs': 0.23331} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.106226] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 825.106397] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed8c2422-60ad-4907-8cc6-6b95b9c75dd7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.112816] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 825.112816] env[61857]: value = "task-950969" [ 825.112816] env[61857]: _type = "Task" [ 825.112816] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.120446] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950969, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.151014] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "93e03e35-15a0-49e7-b1be-09178eabbfda" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.151014] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "93e03e35-15a0-49e7-b1be-09178eabbfda" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.151014] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "93e03e35-15a0-49e7-b1be-09178eabbfda-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.151014] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "93e03e35-15a0-49e7-b1be-09178eabbfda-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.151301] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "93e03e35-15a0-49e7-b1be-09178eabbfda-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.155850] env[61857]: INFO nova.compute.manager [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Terminating instance [ 825.157745] env[61857]: DEBUG nova.compute.manager [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.157949] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 825.158796] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5752b331-edb7-4080-94e6-aaad36a1d3f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.166230] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.166495] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b66958a-44bf-4709-a83b-4c2ecfc37743 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.173113] env[61857]: DEBUG oslo_vmware.api [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 825.173113] env[61857]: value = "task-950970" [ 825.173113] env[61857]: _type = "Task" [ 825.173113] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.181490] env[61857]: DEBUG oslo_vmware.api [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950970, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.215546] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950966, 'name': Destroy_Task, 'duration_secs': 0.784136} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.216360] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Destroyed the VM [ 825.216695] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 825.218604] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7fc71644-1152-4b53-af86-970ff2324ba4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.225586] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 825.225586] env[61857]: value = "task-950971" [ 825.225586] env[61857]: _type = "Task" [ 825.225586] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.236563] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950971, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.294243] env[61857]: DEBUG nova.network.neutron [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Successfully updated port: 0d52a5bd-50d8-41e1-86aa-b1eec00d8702 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.355030] env[61857]: DEBUG nova.compute.manager [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.441390] env[61857]: DEBUG nova.compute.utils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.443934] env[61857]: DEBUG nova.compute.manager [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 825.444213] env[61857]: DEBUG nova.network.neutron [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 825.518664] env[61857]: DEBUG nova.policy [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c23e95cd2584bada69e98ba9e8575ab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '180bf8932fb845c68327ab92ead371d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 825.600553] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950967, 'name': CreateVM_Task, 'duration_secs': 0.524046} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.602833] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 825.603610] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.603799] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.604169] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.605476] env[61857]: DEBUG nova.network.neutron [req-d1509e6b-f81d-4b47-84a6-116aac791f03 req-2f04f7c0-07fc-450f-a25d-1cb9b5d1bfa9 service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Updated VIF entry in instance network info cache for port 3adcb286-015b-4b35-87f4-befacb5b4f3d. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 825.605875] env[61857]: DEBUG nova.network.neutron [req-d1509e6b-f81d-4b47-84a6-116aac791f03 req-2f04f7c0-07fc-450f-a25d-1cb9b5d1bfa9 service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Updating instance_info_cache with network_info: [{"id": "3adcb286-015b-4b35-87f4-befacb5b4f3d", "address": "fa:16:3e:0f:4f:d3", "network": {"id": "a2079895-c2b5-4cb6-9429-3990ae7b910e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-293337110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b21cc6a46cc04f20b942603dc80ec6c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3adcb286-01", "ovs_interfaceid": "3adcb286-015b-4b35-87f4-befacb5b4f3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.607382] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e357b158-feff-4d77-aced-8b4e17183711 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.613694] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 825.613694] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5282b052-10a6-0580-deb9-b372475fdb6c" [ 825.613694] env[61857]: _type = "Task" [ 825.613694] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.626984] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950969, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.629926] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5282b052-10a6-0580-deb9-b372475fdb6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.683539] env[61857]: DEBUG oslo_vmware.api [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950970, 'name': PowerOffVM_Task, 'duration_secs': 0.238333} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.683846] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.684030] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 825.684305] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50d78aa5-902b-4b61-be7b-fbb2d9b2418d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.722613] env[61857]: DEBUG nova.objects.instance [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'flavor' on Instance uuid 71920a62-090d-4df1-937e-87df3b043e28 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.741320] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950971, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.766670] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 825.766670] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 825.766670] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleting the datastore file [datastore1] 93e03e35-15a0-49e7-b1be-09178eabbfda {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.767125] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96622c47-4106-439b-9657-5e104e33f6ee {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.778861] env[61857]: DEBUG oslo_vmware.api [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 825.778861] env[61857]: value = "task-950973" [ 825.778861] env[61857]: _type = "Task" [ 825.778861] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.796793] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "refresh_cache-e5e64477-a1aa-4a4d-91a2-b17d912e09c3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.796964] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "refresh_cache-e5e64477-a1aa-4a4d-91a2-b17d912e09c3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.797148] env[61857]: DEBUG nova.network.neutron [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 825.798422] env[61857]: DEBUG oslo_vmware.api [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950973, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.886409] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.898264] env[61857]: DEBUG nova.network.neutron [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Successfully created port: 71f818ca-5a91-42ed-8825-5141bf75a30e {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.948092] env[61857]: DEBUG nova.compute.manager [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 826.113010] env[61857]: DEBUG oslo_concurrency.lockutils [req-d1509e6b-f81d-4b47-84a6-116aac791f03 req-2f04f7c0-07fc-450f-a25d-1cb9b5d1bfa9 service nova] Releasing lock "refresh_cache-a73f3163-b6fc-4764-b76a-451f823b5808" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.129080] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950969, 'name': PowerOnVM_Task, 'duration_secs': 0.545127} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.129445] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 826.129668] env[61857]: INFO nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Took 9.42 seconds to spawn the instance on the hypervisor. [ 826.129919] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.130717] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebaeaf1a-bb8e-48f6-b927-a8889526184b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.137609] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5282b052-10a6-0580-deb9-b372475fdb6c, 'name': SearchDatastore_Task, 'duration_secs': 0.012821} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.140348] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.140613] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.140874] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.141065] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.141286] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.142578] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ae2937f-55ea-4317-9caa-a0bb72716937 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.154643] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.154872] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 826.155602] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11af13b8-c455-48a9-964c-e6964e1f3c7f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.169307] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 826.169307] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5200cdd0-0576-d01f-e4e8-011e199defbb" [ 826.169307] env[61857]: _type = "Task" [ 826.169307] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.176408] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5200cdd0-0576-d01f-e4e8-011e199defbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.232554] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.232815] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.233092] env[61857]: DEBUG nova.network.neutron [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 826.233349] env[61857]: DEBUG nova.objects.instance [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'info_cache' on Instance uuid 71920a62-090d-4df1-937e-87df3b043e28 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 826.246032] env[61857]: DEBUG oslo_vmware.api [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950971, 'name': RemoveSnapshot_Task, 'duration_secs': 0.828567} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.246032] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 826.246032] env[61857]: INFO nova.compute.manager [None req-9f9aba9a-d840-45fc-a323-5a2d87d1b764 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Took 17.43 seconds to snapshot the instance on the hypervisor. [ 826.291751] env[61857]: DEBUG oslo_vmware.api [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-950973, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166955} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.292034] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.292241] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 826.292417] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 826.292591] env[61857]: INFO nova.compute.manager [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Took 1.13 seconds to destroy the instance on the hypervisor. [ 826.292837] env[61857]: DEBUG oslo.service.loopingcall [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.293048] env[61857]: DEBUG nova.compute.manager [-] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.293146] env[61857]: DEBUG nova.network.neutron [-] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 826.339259] env[61857]: DEBUG nova.network.neutron [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 826.489911] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1317ab-b0f4-46bf-9a94-5adca5df1f16 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.498931] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd8e7ba-f077-4f54-a7fa-38fa3a82d7d3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.530711] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48593add-add8-46f6-b7ff-a7e05e4845fb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.539341] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c79231-63b4-47fd-a91c-eb2d4b10cce3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.545855] env[61857]: DEBUG nova.network.neutron [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Updating instance_info_cache with network_info: [{"id": "0d52a5bd-50d8-41e1-86aa-b1eec00d8702", "address": "fa:16:3e:81:81:2a", "network": {"id": "e7b01bbf-aac5-42cb-b166-da2a3a70a175", "bridge": "br-int", "label": "tempest-ImagesTestJSON-229290731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3688a3ca85044527a84ad1a7b201fbdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d52a5bd-50", "ovs_interfaceid": "0d52a5bd-50d8-41e1-86aa-b1eec00d8702", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.558312] env[61857]: DEBUG nova.compute.provider_tree [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.660776] env[61857]: INFO nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Took 41.71 seconds to build instance. [ 826.677563] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5200cdd0-0576-d01f-e4e8-011e199defbb, 'name': SearchDatastore_Task, 'duration_secs': 0.020772} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.678346] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e876b78-02d7-44ca-8ba9-a6a4559eeef6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.683409] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 826.683409] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]527d0b46-9244-9363-fefa-4b3a1818e500" [ 826.683409] env[61857]: _type = "Task" [ 826.683409] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.691671] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]527d0b46-9244-9363-fefa-4b3a1818e500, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.706585] env[61857]: DEBUG nova.compute.manager [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Received event network-vif-plugged-0d52a5bd-50d8-41e1-86aa-b1eec00d8702 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 826.706797] env[61857]: DEBUG oslo_concurrency.lockutils [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] Acquiring lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.707025] env[61857]: DEBUG oslo_concurrency.lockutils [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] Lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.707205] env[61857]: DEBUG oslo_concurrency.lockutils [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] Lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.707379] env[61857]: DEBUG nova.compute.manager [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] No waiting events found dispatching network-vif-plugged-0d52a5bd-50d8-41e1-86aa-b1eec00d8702 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 826.707549] env[61857]: WARNING nova.compute.manager [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Received unexpected event network-vif-plugged-0d52a5bd-50d8-41e1-86aa-b1eec00d8702 for instance with vm_state building and task_state spawning. [ 826.707718] env[61857]: DEBUG nova.compute.manager [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Received event network-changed-0d52a5bd-50d8-41e1-86aa-b1eec00d8702 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 826.708277] env[61857]: DEBUG nova.compute.manager [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Refreshing instance network info cache due to event network-changed-0d52a5bd-50d8-41e1-86aa-b1eec00d8702. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 826.708277] env[61857]: DEBUG oslo_concurrency.lockutils [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] Acquiring lock "refresh_cache-e5e64477-a1aa-4a4d-91a2-b17d912e09c3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.737174] env[61857]: DEBUG nova.objects.base [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Object Instance<71920a62-090d-4df1-937e-87df3b043e28> lazy-loaded attributes: flavor,info_cache {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 826.963294] env[61857]: DEBUG nova.compute.manager [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 826.989538] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 826.989792] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 826.989964] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 826.990197] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 826.990380] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 826.990540] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 826.990756] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 826.990927] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 826.991105] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 826.991275] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 826.991453] env[61857]: DEBUG nova.virt.hardware [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.992339] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd8b064-a226-43b0-b1fe-9b3705470a13 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.000151] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf896f8-c212-43fb-926c-ba57032d9d70 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.048417] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "refresh_cache-e5e64477-a1aa-4a4d-91a2-b17d912e09c3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.048786] env[61857]: DEBUG nova.compute.manager [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Instance network_info: |[{"id": "0d52a5bd-50d8-41e1-86aa-b1eec00d8702", "address": "fa:16:3e:81:81:2a", "network": {"id": "e7b01bbf-aac5-42cb-b166-da2a3a70a175", "bridge": "br-int", "label": "tempest-ImagesTestJSON-229290731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3688a3ca85044527a84ad1a7b201fbdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d52a5bd-50", "ovs_interfaceid": "0d52a5bd-50d8-41e1-86aa-b1eec00d8702", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 827.049132] env[61857]: DEBUG oslo_concurrency.lockutils [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] Acquired lock "refresh_cache-e5e64477-a1aa-4a4d-91a2-b17d912e09c3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.049332] env[61857]: DEBUG nova.network.neutron [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Refreshing network info cache for port 0d52a5bd-50d8-41e1-86aa-b1eec00d8702 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 827.050685] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:81:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4576b9d4-535c-40aa-b078-246f671f216e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0d52a5bd-50d8-41e1-86aa-b1eec00d8702', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.058204] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Creating folder: Project (3688a3ca85044527a84ad1a7b201fbdd). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 827.061091] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-36e155f2-4c49-4274-839a-11939bd2f4e7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.063483] env[61857]: DEBUG nova.scheduler.client.report [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.075681] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Created folder: Project (3688a3ca85044527a84ad1a7b201fbdd) in parent group-v214027. [ 827.075681] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Creating folder: Instances. Parent ref: group-v214111. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 827.075791] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-68b23f11-463b-4f82-a59f-32a70cf3aa78 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.084679] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Created folder: Instances in parent group-v214111. [ 827.085632] env[61857]: DEBUG oslo.service.loopingcall [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.085632] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 827.085632] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49115c49-cbee-4e25-8f8f-473c2d63ed60 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.107053] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.107053] env[61857]: value = "task-950976" [ 827.107053] env[61857]: _type = "Task" [ 827.107053] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.116869] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950976, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.140806] env[61857]: DEBUG nova.network.neutron [-] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.163045] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.589s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.196740] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]527d0b46-9244-9363-fefa-4b3a1818e500, 'name': SearchDatastore_Task, 'duration_secs': 0.050677} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.197091] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.197424] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] a73f3163-b6fc-4764-b76a-451f823b5808/a73f3163-b6fc-4764-b76a-451f823b5808.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 827.197728] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e59ed8e4-8c5c-4c47-b4ca-f025f48e7e0a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.206638] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 827.206638] env[61857]: value = "task-950977" [ 827.206638] env[61857]: _type = "Task" [ 827.206638] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.215836] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950977, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.341975] env[61857]: DEBUG nova.network.neutron [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Updated VIF entry in instance network info cache for port 0d52a5bd-50d8-41e1-86aa-b1eec00d8702. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 827.342170] env[61857]: DEBUG nova.network.neutron [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Updating instance_info_cache with network_info: [{"id": "0d52a5bd-50d8-41e1-86aa-b1eec00d8702", "address": "fa:16:3e:81:81:2a", "network": {"id": "e7b01bbf-aac5-42cb-b166-da2a3a70a175", "bridge": "br-int", "label": "tempest-ImagesTestJSON-229290731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3688a3ca85044527a84ad1a7b201fbdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d52a5bd-50", "ovs_interfaceid": "0d52a5bd-50d8-41e1-86aa-b1eec00d8702", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.510059] env[61857]: DEBUG nova.network.neutron [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance_info_cache with network_info: [{"id": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "address": "fa:16:3e:32:96:34", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c3f890-dd", "ovs_interfaceid": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.568514] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.630s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.569126] env[61857]: DEBUG nova.compute.manager [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.572577] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 29.533s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.622451] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950976, 'name': CreateVM_Task, 'duration_secs': 0.359083} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.622710] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 827.623535] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.623775] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.624196] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 827.624523] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e942c80-b825-479d-a231-887c3cd574e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.630341] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 827.630341] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ccee8d-966a-69fa-01aa-4756cfdff252" [ 827.630341] env[61857]: _type = "Task" [ 827.630341] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.640012] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ccee8d-966a-69fa-01aa-4756cfdff252, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.643858] env[61857]: INFO nova.compute.manager [-] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Took 1.35 seconds to deallocate network for instance. [ 827.666790] env[61857]: DEBUG nova.compute.manager [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.718934] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950977, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.774819] env[61857]: DEBUG nova.network.neutron [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Successfully updated port: 71f818ca-5a91-42ed-8825-5141bf75a30e {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 827.846182] env[61857]: DEBUG oslo_concurrency.lockutils [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] Releasing lock "refresh_cache-e5e64477-a1aa-4a4d-91a2-b17d912e09c3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.846182] env[61857]: DEBUG nova.compute.manager [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Received event network-vif-deleted-7ae4b256-08e6-483a-a76a-1c1bcc44e67d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 827.846528] env[61857]: INFO nova.compute.manager [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Neutron deleted interface 7ae4b256-08e6-483a-a76a-1c1bcc44e67d; detaching it from the instance and deleting it from the info cache [ 827.846528] env[61857]: DEBUG nova.network.neutron [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.013418] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.081719] env[61857]: DEBUG nova.compute.utils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.086839] env[61857]: INFO nova.compute.claims [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.091225] env[61857]: DEBUG nova.compute.manager [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Not allocating networking since 'none' was specified. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 828.142462] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ccee8d-966a-69fa-01aa-4756cfdff252, 'name': SearchDatastore_Task, 'duration_secs': 0.064847} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.142924] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.143221] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.143479] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.143677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.143829] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.144119] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1317b44-45bf-4224-bb58-e43313da1730 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.152661] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.152855] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 828.154222] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.154462] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d43af58b-efc7-4dcd-bd86-caaa5506f58d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.161055] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 828.161055] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e14082-4675-8632-6eb1-3d7910d0ae27" [ 828.161055] env[61857]: _type = "Task" [ 828.161055] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.169808] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e14082-4675-8632-6eb1-3d7910d0ae27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.195764] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.218303] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950977, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.721485} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.218611] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] a73f3163-b6fc-4764-b76a-451f823b5808/a73f3163-b6fc-4764-b76a-451f823b5808.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 828.218834] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.219109] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e165af89-ce80-4cc0-a6f8-4c1838247872 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.225743] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 828.225743] env[61857]: value = "task-950978" [ 828.225743] env[61857]: _type = "Task" [ 828.225743] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.233962] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950978, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.278206] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "refresh_cache-16b447b5-426a-4478-9d44-ae32b41dee50" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.278360] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquired lock "refresh_cache-16b447b5-426a-4478-9d44-ae32b41dee50" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.278515] env[61857]: DEBUG nova.network.neutron [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.349321] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9df6bc96-5a2e-4ae5-886e-0541a4888e96 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.359517] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61517487-b17c-46b7-91d3-0413644fe966 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.389700] env[61857]: DEBUG nova.compute.manager [req-aa1754b4-7906-4cd0-b8c6-a39afdebcc41 req-412b300e-b40b-41fc-b407-4268307252d6 service nova] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Detach interface failed, port_id=7ae4b256-08e6-483a-a76a-1c1bcc44e67d, reason: Instance 93e03e35-15a0-49e7-b1be-09178eabbfda could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 828.516826] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 828.517233] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cfcf7e9a-c289-4d63-a9bc-66dc6c10c14c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.524280] env[61857]: DEBUG oslo_vmware.api [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 828.524280] env[61857]: value = "task-950979" [ 828.524280] env[61857]: _type = "Task" [ 828.524280] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.532609] env[61857]: DEBUG oslo_vmware.api [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950979, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.593169] env[61857]: INFO nova.compute.resource_tracker [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating resource usage from migration f0272d66-8a8c-48a8-aad1-69a9053fe892 [ 828.596438] env[61857]: DEBUG nova.compute.manager [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.621302] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.621302] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.672442] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e14082-4675-8632-6eb1-3d7910d0ae27, 'name': SearchDatastore_Task, 'duration_secs': 0.009894} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.673268] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdda35b5-2fbe-4457-abd1-22027b072d42 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.678456] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 828.678456] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]528f6af2-f9f4-efd6-88ed-10afb27872ac" [ 828.678456] env[61857]: _type = "Task" [ 828.678456] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.681672] env[61857]: DEBUG nova.compute.manager [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.685751] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1e5527-f389-4bd4-b71c-1c313bcdc76d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.700447] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528f6af2-f9f4-efd6-88ed-10afb27872ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.736413] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950978, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074059} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.736701] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.737526] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f92308a-550d-4461-8de6-86d06bee5976 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.763068] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] a73f3163-b6fc-4764-b76a-451f823b5808/a73f3163-b6fc-4764-b76a-451f823b5808.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.763432] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a87e7077-38b3-467a-b9c6-953ff20dd9ea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.788637] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 828.788637] env[61857]: value = "task-950980" [ 828.788637] env[61857]: _type = "Task" [ 828.788637] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.796819] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950980, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.812174] env[61857]: DEBUG nova.network.neutron [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 828.821532] env[61857]: DEBUG nova.compute.manager [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Received event network-vif-plugged-71f818ca-5a91-42ed-8825-5141bf75a30e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.821777] env[61857]: DEBUG oslo_concurrency.lockutils [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] Acquiring lock "16b447b5-426a-4478-9d44-ae32b41dee50-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.821989] env[61857]: DEBUG oslo_concurrency.lockutils [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] Lock "16b447b5-426a-4478-9d44-ae32b41dee50-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.822185] env[61857]: DEBUG oslo_concurrency.lockutils [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] Lock "16b447b5-426a-4478-9d44-ae32b41dee50-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.822374] env[61857]: DEBUG nova.compute.manager [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] No waiting events found dispatching network-vif-plugged-71f818ca-5a91-42ed-8825-5141bf75a30e {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.822546] env[61857]: WARNING nova.compute.manager [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Received unexpected event network-vif-plugged-71f818ca-5a91-42ed-8825-5141bf75a30e for instance with vm_state building and task_state spawning. [ 828.822708] env[61857]: DEBUG nova.compute.manager [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Received event network-changed-71f818ca-5a91-42ed-8825-5141bf75a30e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.822867] env[61857]: DEBUG nova.compute.manager [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Refreshing instance network info cache due to event network-changed-71f818ca-5a91-42ed-8825-5141bf75a30e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 828.823050] env[61857]: DEBUG oslo_concurrency.lockutils [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] Acquiring lock "refresh_cache-16b447b5-426a-4478-9d44-ae32b41dee50" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.965889] env[61857]: DEBUG nova.network.neutron [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Updating instance_info_cache with network_info: [{"id": "71f818ca-5a91-42ed-8825-5141bf75a30e", "address": "fa:16:3e:80:2f:1f", "network": {"id": "b4be0ad1-dff9-4093-8710-71550c9d3525", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1005551508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180bf8932fb845c68327ab92ead371d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2c68e7-b690-42e2-9491-c3f9357cc66a", "external-id": "nsx-vlan-transportzone-321", "segmentation_id": 321, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71f818ca-5a", "ovs_interfaceid": "71f818ca-5a91-42ed-8825-5141bf75a30e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.037113] env[61857]: DEBUG oslo_vmware.api [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950979, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.049792] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c825231d-503c-4faf-9745-0c13f1d80e63 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.056756] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faaf2cc1-9881-4e95-b0df-986a857fea8c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.088947] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12631fca-2af9-4747-9c37-3eef52385bfc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.097139] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f54658-4d57-4950-be1a-e4265ce130b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.114525] env[61857]: DEBUG nova.compute.provider_tree [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.124170] env[61857]: DEBUG nova.compute.utils [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.191816] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528f6af2-f9f4-efd6-88ed-10afb27872ac, 'name': SearchDatastore_Task, 'duration_secs': 0.025877} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.192121] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.192398] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] e5e64477-a1aa-4a4d-91a2-b17d912e09c3/e5e64477-a1aa-4a4d-91a2-b17d912e09c3.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 829.192683] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68b2fa84-b40f-4a7f-9e66-2960205add57 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.199109] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 829.199109] env[61857]: value = "task-950981" [ 829.199109] env[61857]: _type = "Task" [ 829.199109] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.202795] env[61857]: INFO nova.compute.manager [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] instance snapshotting [ 829.205682] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c66b694-d567-4604-a20e-8d80adec5e21 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.210963] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.227597] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a991b29-3947-4387-8796-164e8a5cbb75 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.300308] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950980, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.469068] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Releasing lock "refresh_cache-16b447b5-426a-4478-9d44-ae32b41dee50" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.469438] env[61857]: DEBUG nova.compute.manager [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Instance network_info: |[{"id": "71f818ca-5a91-42ed-8825-5141bf75a30e", "address": "fa:16:3e:80:2f:1f", "network": {"id": "b4be0ad1-dff9-4093-8710-71550c9d3525", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1005551508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180bf8932fb845c68327ab92ead371d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2c68e7-b690-42e2-9491-c3f9357cc66a", "external-id": "nsx-vlan-transportzone-321", "segmentation_id": 321, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71f818ca-5a", "ovs_interfaceid": "71f818ca-5a91-42ed-8825-5141bf75a30e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 829.469762] env[61857]: DEBUG oslo_concurrency.lockutils [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] Acquired lock "refresh_cache-16b447b5-426a-4478-9d44-ae32b41dee50" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.469950] env[61857]: DEBUG nova.network.neutron [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Refreshing network info cache for port 71f818ca-5a91-42ed-8825-5141bf75a30e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 829.471432] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:2f:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb2c68e7-b690-42e2-9491-c3f9357cc66a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '71f818ca-5a91-42ed-8825-5141bf75a30e', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.478895] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Creating folder: Project (180bf8932fb845c68327ab92ead371d7). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 829.482358] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46278ae5-d223-4d04-a242-02d3f849a918 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.493052] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Created folder: Project (180bf8932fb845c68327ab92ead371d7) in parent group-v214027. [ 829.493155] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Creating folder: Instances. Parent ref: group-v214114. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 829.493384] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d7ed0e10-975a-4b29-aba3-307f734cd705 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.502246] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Created folder: Instances in parent group-v214114. [ 829.502528] env[61857]: DEBUG oslo.service.loopingcall [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.502756] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 829.502967] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b0d5d58-5a4a-46a6-823d-6a9196ea2ce6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.523379] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.523379] env[61857]: value = "task-950984" [ 829.523379] env[61857]: _type = "Task" [ 829.523379] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.533105] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950984, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.537812] env[61857]: DEBUG oslo_vmware.api [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-950979, 'name': PowerOnVM_Task, 'duration_secs': 0.527583} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.538108] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 829.538374] env[61857]: DEBUG nova.compute.manager [None req-e0aab466-01d5-4bb1-9512-ca307696b07c tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.539103] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afaacc31-ee2d-42b1-b6f8-543411f5e6fa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.606792] env[61857]: DEBUG nova.compute.manager [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.617387] env[61857]: DEBUG nova.scheduler.client.report [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.628923] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.631776] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.632140] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.632421] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.632658] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.632818] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.633261] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.633261] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.633437] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.633625] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.633798] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.633992] env[61857]: DEBUG nova.virt.hardware [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.635410] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa369a7-1b67-47da-beb2-bb62fb5f1d66 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.645631] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e0e11a-f77e-45a8-b0cd-7db8477fc9e2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.663293] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.671022] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Creating folder: Project (18d87d345abe419d857e19dae16d9cf1). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 829.671022] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5fc81b9-432a-4800-9fe7-a71015ab28c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.685178] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Created folder: Project (18d87d345abe419d857e19dae16d9cf1) in parent group-v214027. [ 829.685178] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Creating folder: Instances. Parent ref: group-v214117. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 829.685178] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98a2f3c4-4c34-4161-9544-50976457b8b4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.696024] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Created folder: Instances in parent group-v214117. [ 829.696024] env[61857]: DEBUG oslo.service.loopingcall [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.696024] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 829.696024] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9993744c-f83c-4abe-8bd6-4925261f1903 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.721616] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950981, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.723803] env[61857]: DEBUG nova.network.neutron [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Updated VIF entry in instance network info cache for port 71f818ca-5a91-42ed-8825-5141bf75a30e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 829.724276] env[61857]: DEBUG nova.network.neutron [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Updating instance_info_cache with network_info: [{"id": "71f818ca-5a91-42ed-8825-5141bf75a30e", "address": "fa:16:3e:80:2f:1f", "network": {"id": "b4be0ad1-dff9-4093-8710-71550c9d3525", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1005551508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180bf8932fb845c68327ab92ead371d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2c68e7-b690-42e2-9491-c3f9357cc66a", "external-id": "nsx-vlan-transportzone-321", "segmentation_id": 321, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71f818ca-5a", "ovs_interfaceid": "71f818ca-5a91-42ed-8825-5141bf75a30e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.725509] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.725509] env[61857]: value = "task-950987" [ 829.725509] env[61857]: _type = "Task" [ 829.725509] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.734466] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950987, 'name': CreateVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.737448] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 829.738173] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-83176d4a-116b-4704-8f69-5b5f8f0e27af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.746960] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 829.746960] env[61857]: value = "task-950988" [ 829.746960] env[61857]: _type = "Task" [ 829.746960] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.756532] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950988, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.802797] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950980, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.033424] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950984, 'name': CreateVM_Task, 'duration_secs': 0.37554} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.033613] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 830.034349] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.034781] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.035149] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.035413] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4d3faac-20f7-4d09-a37c-395f6abd354d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.040277] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 830.040277] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5252e36f-b5a1-fc1e-77d3-301497232967" [ 830.040277] env[61857]: _type = "Task" [ 830.040277] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.052087] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5252e36f-b5a1-fc1e-77d3-301497232967, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.125365] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.553s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.125564] env[61857]: INFO nova.compute.manager [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Migrating [ 830.125800] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.125952] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.127294] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.781s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.128845] env[61857]: INFO nova.compute.claims [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.132326] env[61857]: INFO nova.compute.rpcapi [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 830.132625] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.223031] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950981, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.738002} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.223429] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] e5e64477-a1aa-4a4d-91a2-b17d912e09c3/e5e64477-a1aa-4a4d-91a2-b17d912e09c3.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 830.223757] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.224092] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5ac84d9-7420-4224-aff0-fe10940a6904 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.226632] env[61857]: DEBUG oslo_concurrency.lockutils [req-b436c7b1-986a-465f-ae0b-cfc267a0f92d req-685e2f3d-be40-495d-93e3-141b283b71ee service nova] Releasing lock "refresh_cache-16b447b5-426a-4478-9d44-ae32b41dee50" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.236509] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-950987, 'name': CreateVM_Task, 'duration_secs': 0.484205} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.237871] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 830.238234] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 830.238234] env[61857]: value = "task-950989" [ 830.238234] env[61857]: _type = "Task" [ 830.238234] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.238724] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.256118] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950988, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.302144] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950980, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.550938] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5252e36f-b5a1-fc1e-77d3-301497232967, 'name': SearchDatastore_Task, 'duration_secs': 0.012353} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.551291] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.551545] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.551804] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.551959] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.552231] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.552631] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.553205] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.553444] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc15b753-2add-4baa-abb2-3bc169ddad2e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.555402] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c62f1d90-9c54-4b9a-8eae-97f0bc34dd43 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.560821] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 830.560821] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5244432b-7900-15a3-6b5b-b9eec8914027" [ 830.560821] env[61857]: _type = "Task" [ 830.560821] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.566262] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.566490] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 830.569866] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-487c8299-d3c8-4290-afeb-3b796e44a0e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.572103] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5244432b-7900-15a3-6b5b-b9eec8914027, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.575014] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 830.575014] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]528309bc-7b94-082d-3c6a-709ecf5f2d16" [ 830.575014] env[61857]: _type = "Task" [ 830.575014] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.582608] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528309bc-7b94-082d-3c6a-709ecf5f2d16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.647905] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.648211] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.648522] env[61857]: DEBUG nova.network.neutron [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.698506] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.698792] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.699045] env[61857]: INFO nova.compute.manager [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Attaching volume 8e4ac73a-5a8b-4540-927e-a4d19d2affd9 to /dev/sdb [ 830.732687] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96737192-f949-481c-80c3-ba5fb3c1b5ce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.743369] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e73ab72-00af-4403-a790-508337fd9069 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.750899] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950989, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077197} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.753910] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 830.758093] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa315a1-848a-4ef3-a6f0-3615a1b3005b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.760598] env[61857]: DEBUG nova.virt.block_device [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Updating existing volume attachment record: cf43c7f1-7725-474a-aba0-b27e736e9dd4 {{(pid=61857) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 830.767883] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950988, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.785462] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] e5e64477-a1aa-4a4d-91a2-b17d912e09c3/e5e64477-a1aa-4a4d-91a2-b17d912e09c3.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.785798] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d206947-47e6-4b66-bb66-d314fa013125 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.811090] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950980, 'name': ReconfigVM_Task, 'duration_secs': 1.577368} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.811342] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Reconfigured VM instance instance-00000040 to attach disk [datastore1] a73f3163-b6fc-4764-b76a-451f823b5808/a73f3163-b6fc-4764-b76a-451f823b5808.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.812157] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 830.812157] env[61857]: value = "task-950990" [ 830.812157] env[61857]: _type = "Task" [ 830.812157] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.812382] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ed5ed2c-f31f-47ae-9432-c9c5ae58f4ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.823279] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950990, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.824763] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 830.824763] env[61857]: value = "task-950991" [ 830.824763] env[61857]: _type = "Task" [ 830.824763] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.833601] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950991, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.070928] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5244432b-7900-15a3-6b5b-b9eec8914027, 'name': SearchDatastore_Task, 'duration_secs': 0.04417} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.071291] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.071541] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.071760] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.086062] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528309bc-7b94-082d-3c6a-709ecf5f2d16, 'name': SearchDatastore_Task, 'duration_secs': 0.038957} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.086539] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bcca03d-4593-482c-9fbc-d63629235e53 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.092187] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 831.092187] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e28e80-6e44-62c9-01e7-5465d79a0254" [ 831.092187] env[61857]: _type = "Task" [ 831.092187] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.102593] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e28e80-6e44-62c9-01e7-5465d79a0254, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.264358] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950988, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.323767] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950990, 'name': ReconfigVM_Task, 'duration_secs': 0.293101} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.327021] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Reconfigured VM instance instance-00000041 to attach disk [datastore1] e5e64477-a1aa-4a4d-91a2-b17d912e09c3/e5e64477-a1aa-4a4d-91a2-b17d912e09c3.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.328105] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4a53200-3086-440c-ac27-e8e8722e44b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.342537] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950991, 'name': Rename_Task, 'duration_secs': 0.173242} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.347458] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 831.347993] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 831.347993] env[61857]: value = "task-950995" [ 831.347993] env[61857]: _type = "Task" [ 831.347993] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.350622] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-120c9db5-10e4-449b-b8b1-b9e27f2aeb4f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.362445] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950995, 'name': Rename_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.368360] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 831.368360] env[61857]: value = "task-950996" [ 831.368360] env[61857]: _type = "Task" [ 831.368360] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.380349] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950996, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.483187] env[61857]: DEBUG nova.network.neutron [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance_info_cache with network_info: [{"id": "9d8efcbb-1103-49f6-a52f-219e1682d275", "address": "fa:16:3e:54:79:1a", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d8efcbb-11", "ovs_interfaceid": "9d8efcbb-1103-49f6-a52f-219e1682d275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.602161] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e28e80-6e44-62c9-01e7-5465d79a0254, 'name': SearchDatastore_Task, 'duration_secs': 0.016913} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.603387] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.603658] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 16b447b5-426a-4478-9d44-ae32b41dee50/16b447b5-426a-4478-9d44-ae32b41dee50.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 831.604450] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7256c430-27c1-4659-a891-ae2debc5c6ea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.607044] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.607259] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.607481] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7b45263-28ad-497d-897e-e51efd5909d4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.609548] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-67be2346-6236-4d7f-a031-5c283eb90b0d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.616169] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bab87bf-cc08-43b2-b0b9-aec0247d7d6a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.620340] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.620534] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 831.621506] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21622d7b-98c3-46c2-9a5e-b7368d602ba3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.650294] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbccfd15-0e46-43e5-89bc-020bcef969cb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.653214] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 831.653214] env[61857]: value = "task-950997" [ 831.653214] env[61857]: _type = "Task" [ 831.653214] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.654631] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 831.654631] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52bed9bf-7886-0a14-a906-cca122ca0c18" [ 831.654631] env[61857]: _type = "Task" [ 831.654631] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.663910] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5fb92d5-1c67-4191-b001-02956bf7f480 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.670656] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-950997, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.675010] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52bed9bf-7886-0a14-a906-cca122ca0c18, 'name': SearchDatastore_Task, 'duration_secs': 0.010506} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.685658] env[61857]: DEBUG nova.compute.provider_tree [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.686965] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a01f7de9-8c79-4b1f-b5df-403d2b065d12 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.693364] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 831.693364] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52633259-6b34-3eb4-bc7a-24c31eeccc44" [ 831.693364] env[61857]: _type = "Task" [ 831.693364] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.702966] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52633259-6b34-3eb4-bc7a-24c31eeccc44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.762032] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-950988, 'name': CreateSnapshot_Task, 'duration_secs': 1.675076} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.762032] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 831.762441] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fcd92a1-92ca-4e89-b50a-7b6b4d221183 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.861901] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950995, 'name': Rename_Task, 'duration_secs': 0.255192} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.862248] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 831.862600] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f3cd425-c36b-4988-af02-ec131bcaac34 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.869355] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 831.869355] env[61857]: value = "task-950998" [ 831.869355] env[61857]: _type = "Task" [ 831.869355] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.881114] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950998, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.885280] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950996, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.986380] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.165157] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-950997, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.192174] env[61857]: DEBUG nova.scheduler.client.report [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.208119] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52633259-6b34-3eb4-bc7a-24c31eeccc44, 'name': SearchDatastore_Task, 'duration_secs': 0.010375} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.208505] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.209193] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] ece27044-3123-4685-a4ab-30eeb10337d1/ece27044-3123-4685-a4ab-30eeb10337d1.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 832.209193] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5c26b59-1236-4395-90aa-4a846ef8ca6f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.218795] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 832.218795] env[61857]: value = "task-950999" [ 832.218795] env[61857]: _type = "Task" [ 832.218795] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.229843] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-950999, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.281028] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 832.281755] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c83da308-8e74-46ef-86f2-7703c8557c0c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.291074] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 832.291074] env[61857]: value = "task-951000" [ 832.291074] env[61857]: _type = "Task" [ 832.291074] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.299493] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951000, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.382847] env[61857]: DEBUG oslo_vmware.api [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-950996, 'name': PowerOnVM_Task, 'duration_secs': 0.638163} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.386483] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 832.386786] env[61857]: INFO nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Took 10.84 seconds to spawn the instance on the hypervisor. [ 832.387134] env[61857]: DEBUG nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.387458] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950998, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.388257] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0087fce5-93b6-438d-8bd1-b450de9dfafc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.666017] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-950997, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685922} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.666312] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 16b447b5-426a-4478-9d44-ae32b41dee50/16b447b5-426a-4478-9d44-ae32b41dee50.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 832.666525] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.666840] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cbab9790-56d8-41a1-8a83-c64746e8593e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.674733] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 832.674733] env[61857]: value = "task-951001" [ 832.674733] env[61857]: _type = "Task" [ 832.674733] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.683416] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951001, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.701485] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.702101] env[61857]: DEBUG nova.compute.manager [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.704781] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.973s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.705014] env[61857]: DEBUG nova.objects.instance [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lazy-loading 'resources' on Instance uuid 2c187f64-8a36-4dfd-94e3-8ea944dbac24 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 832.728528] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-950999, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.802164] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951000, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.883355] env[61857]: DEBUG oslo_vmware.api [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-950998, 'name': PowerOnVM_Task, 'duration_secs': 0.526974} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.883633] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 832.884434] env[61857]: INFO nova.compute.manager [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Took 8.56 seconds to spawn the instance on the hypervisor. [ 832.884434] env[61857]: DEBUG nova.compute.manager [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.884869] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acccdf86-9f91-42f1-a995-6c6b94b2e79a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.908328] env[61857]: INFO nova.compute.manager [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Took 44.34 seconds to build instance. [ 833.185355] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951001, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.275912} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.185789] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.186877] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adf41fd-0c64-47b6-884d-7a21241e4e0e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.213866] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 16b447b5-426a-4478-9d44-ae32b41dee50/16b447b5-426a-4478-9d44-ae32b41dee50.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.215931] env[61857]: DEBUG nova.compute.utils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.221216] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09cf7ef2-dfee-49fc-ad65-74caca4618bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.236860] env[61857]: DEBUG nova.compute.manager [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.237126] env[61857]: DEBUG nova.network.neutron [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 833.249206] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-950999, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.991989} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.250283] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] ece27044-3123-4685-a4ab-30eeb10337d1/ece27044-3123-4685-a4ab-30eeb10337d1.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 833.250516] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 833.250833] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 833.250833] env[61857]: value = "task-951003" [ 833.250833] env[61857]: _type = "Task" [ 833.250833] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.251460] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c10d5ba5-4951-4265-9ccb-541c23766555 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.272027] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.272027] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 833.272027] env[61857]: value = "task-951004" [ 833.272027] env[61857]: _type = "Task" [ 833.272027] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.287137] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951004, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.304060] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951000, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.341189] env[61857]: DEBUG nova.policy [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9eee4d78af2540a8a2dfdcafa652a7ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f520132b3ea465aa25f665b8a7750bd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.405423] env[61857]: INFO nova.compute.manager [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Took 43.72 seconds to build instance. [ 833.409290] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fadca9b1-4eb6-45ab-84f1-bd3f043c16e4 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "a73f3163-b6fc-4764-b76a-451f823b5808" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.801s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.506354] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e95f4693-8483-42e5-9bd1-f9adc882d4a4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.530378] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance '30d905ed-831d-44ae-807c-062de9a7e9fb' progress to 0 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 833.716187] env[61857]: DEBUG oslo_concurrency.lockutils [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.716187] env[61857]: DEBUG oslo_concurrency.lockutils [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.716187] env[61857]: DEBUG oslo_concurrency.lockutils [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.716187] env[61857]: DEBUG oslo_concurrency.lockutils [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.716635] env[61857]: DEBUG oslo_concurrency.lockutils [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.721814] env[61857]: DEBUG nova.compute.manager [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.725618] env[61857]: INFO nova.compute.manager [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Terminating instance [ 833.732020] env[61857]: DEBUG nova.compute.manager [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 833.732020] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 833.732020] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f96608-e0c8-463e-9987-ab1f47b0771b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.737989] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 833.740930] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b015a9ef-53ce-452c-ad03-9983663ec835 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.749822] env[61857]: DEBUG oslo_vmware.api [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 833.749822] env[61857]: value = "task-951005" [ 833.749822] env[61857]: _type = "Task" [ 833.749822] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.763465] env[61857]: DEBUG oslo_vmware.api [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-951005, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.772933] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951003, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.774253] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "a73f3163-b6fc-4764-b76a-451f823b5808" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.774573] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "a73f3163-b6fc-4764-b76a-451f823b5808" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.774640] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "a73f3163-b6fc-4764-b76a-451f823b5808-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.774814] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "a73f3163-b6fc-4764-b76a-451f823b5808-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.774990] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "a73f3163-b6fc-4764-b76a-451f823b5808-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.781565] env[61857]: INFO nova.compute.manager [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Terminating instance [ 833.789043] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d74689-7cec-4eb9-ba95-1b2816ca207b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.792322] env[61857]: DEBUG nova.compute.manager [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 833.792444] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 833.793983] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c817a169-a6f2-4594-8017-9efc8f0ac88a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.805234] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951004, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068563} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.807953] env[61857]: DEBUG nova.network.neutron [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Successfully created port: 68fc598e-90c4-40b6-b89a-8e4fe65dd7dd {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.809781] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.810930] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-072c25b4-e8c6-497a-9f40-477254abe2f7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.814033] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02afcf21-e2fa-4357-af63-3092f403a433 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.823102] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 833.823690] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951000, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.825245] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7539bfd-76d3-49e3-b581-5d212772eff3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.843733] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] ece27044-3123-4685-a4ab-30eeb10337d1/ece27044-3123-4685-a4ab-30eeb10337d1.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.847146] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-362aac0f-999f-4895-9ab1-cc7698f06b74 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.861780] env[61857]: DEBUG oslo_vmware.api [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 833.861780] env[61857]: value = "task-951006" [ 833.861780] env[61857]: _type = "Task" [ 833.861780] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.896495] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 833.896495] env[61857]: value = "task-951007" [ 833.896495] env[61857]: _type = "Task" [ 833.896495] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.898264] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3294774c-0c67-41ad-9565-00cc2b998d9d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.907039] env[61857]: DEBUG oslo_vmware.api [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-951006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.907393] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dedb4b90-d012-4d69-a29d-ec71ca235c5c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.676s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.915849] env[61857]: DEBUG nova.compute.manager [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 833.918770] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951007, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.920231] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3867de1c-16a2-4be5-9ba5-e558dc397bf7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.936548] env[61857]: DEBUG nova.compute.provider_tree [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.040451] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 834.040874] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d9496da-4352-4fb6-a3f4-154636a2d79c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.048118] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 834.048118] env[61857]: value = "task-951008" [ 834.048118] env[61857]: _type = "Task" [ 834.048118] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.057475] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951008, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.261565] env[61857]: DEBUG oslo_vmware.api [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-951005, 'name': PowerOffVM_Task, 'duration_secs': 0.485093} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.262247] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 834.262428] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 834.262679] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7221d2af-5d21-449c-b07e-befd5e6dedb9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.266822] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951003, 'name': ReconfigVM_Task, 'duration_secs': 0.766054} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.267987] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 16b447b5-426a-4478-9d44-ae32b41dee50/16b447b5-426a-4478-9d44-ae32b41dee50.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.268686] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4bf2d1c3-481e-40ff-b68d-ed42709c7a02 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.276156] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 834.276156] env[61857]: value = "task-951010" [ 834.276156] env[61857]: _type = "Task" [ 834.276156] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.287314] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951010, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.302313] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951000, 'name': CloneVM_Task, 'duration_secs': 1.885347} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.302613] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Created linked-clone VM from snapshot [ 834.303387] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2379cc99-b5b5-47ec-8507-e5b4b6dbc81a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.311832] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Uploading image 31bb2f90-e5f9-4882-a312-c73ca6d91deb {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 834.326859] env[61857]: DEBUG nova.network.neutron [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Successfully created port: b19c85c5-d538-48ac-b311-493be32a9b2e {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.342344] env[61857]: DEBUG oslo_vmware.rw_handles [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 834.342344] env[61857]: value = "vm-214123" [ 834.342344] env[61857]: _type = "VirtualMachine" [ 834.342344] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 834.343567] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-fffa5abd-4d23-4664-a149-4eff2d8a5ff2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.351281] env[61857]: DEBUG oslo_vmware.rw_handles [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lease: (returnval){ [ 834.351281] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525fa053-74a1-9062-c26e-e2e3a5eb66ca" [ 834.351281] env[61857]: _type = "HttpNfcLease" [ 834.351281] env[61857]: } obtained for exporting VM: (result){ [ 834.351281] env[61857]: value = "vm-214123" [ 834.351281] env[61857]: _type = "VirtualMachine" [ 834.351281] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 834.351569] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the lease: (returnval){ [ 834.351569] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525fa053-74a1-9062-c26e-e2e3a5eb66ca" [ 834.351569] env[61857]: _type = "HttpNfcLease" [ 834.351569] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 834.361536] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 834.361536] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525fa053-74a1-9062-c26e-e2e3a5eb66ca" [ 834.361536] env[61857]: _type = "HttpNfcLease" [ 834.361536] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 834.372934] env[61857]: DEBUG oslo_vmware.api [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-951006, 'name': PowerOffVM_Task, 'duration_secs': 0.253093} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.373220] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 834.373392] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 834.373644] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5cb9304-7914-4355-a6ac-31e4c947cde1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.406994] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951007, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.410594] env[61857]: DEBUG nova.compute.manager [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 834.424864] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 834.425709] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 834.425709] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Deleting the datastore file [datastore1] bf2f46d4-5c6e-4493-b087-fe5a316b8c88 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.428493] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e629237-798f-4e33-b98c-f27af769e9aa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.436706] env[61857]: DEBUG oslo_vmware.api [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 834.436706] env[61857]: value = "task-951013" [ 834.436706] env[61857]: _type = "Task" [ 834.436706] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.441361] env[61857]: DEBUG nova.scheduler.client.report [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.446480] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 834.446997] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 834.446997] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Deleting the datastore file [datastore1] a73f3163-b6fc-4764-b76a-451f823b5808 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.447893] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1dd1f6ec-8ced-44ee-9a94-3a47b621c0dd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.457129] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.458483] env[61857]: DEBUG oslo_vmware.api [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-951013, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.462999] env[61857]: DEBUG oslo_vmware.api [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for the task: (returnval){ [ 834.462999] env[61857]: value = "task-951014" [ 834.462999] env[61857]: _type = "Task" [ 834.462999] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.471634] env[61857]: DEBUG oslo_vmware.api [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-951014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.557781] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951008, 'name': PowerOffVM_Task, 'duration_secs': 0.323937} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.558085] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 834.558295] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance '30d905ed-831d-44ae-807c-062de9a7e9fb' progress to 17 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 834.734836] env[61857]: DEBUG nova.compute.manager [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.755583] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.755852] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.756030] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.756234] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.756391] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.756547] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.756759] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.756920] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.757102] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.757274] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.757459] env[61857]: DEBUG nova.virt.hardware [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.758356] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a915be9a-6522-4fa6-bf18-3d39f5319be8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.766279] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35502034-1b7b-4398-8b91-3eb4546473b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.789073] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951010, 'name': Rename_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.859411] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 834.859411] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525fa053-74a1-9062-c26e-e2e3a5eb66ca" [ 834.859411] env[61857]: _type = "HttpNfcLease" [ 834.859411] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 834.859998] env[61857]: DEBUG oslo_vmware.rw_handles [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 834.859998] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525fa053-74a1-9062-c26e-e2e3a5eb66ca" [ 834.859998] env[61857]: _type = "HttpNfcLease" [ 834.859998] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 834.860802] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef266f0-6995-47bf-a1f2-3b133f863627 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.868037] env[61857]: DEBUG oslo_vmware.rw_handles [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a91a3e-0e61-e244-9b8c-262e7e7a0092/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 834.868231] env[61857]: DEBUG oslo_vmware.rw_handles [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a91a3e-0e61-e244-9b8c-262e7e7a0092/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 834.951747] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.247s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.953791] env[61857]: DEBUG oslo_vmware.api [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-951013, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.366872} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.954024] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951007, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.955519] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.955519] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.112s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.959017] env[61857]: INFO nova.compute.claims [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.959277] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.959467] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 834.959644] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 834.959813] env[61857]: INFO nova.compute.manager [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Took 1.23 seconds to destroy the instance on the hypervisor. [ 834.960063] env[61857]: DEBUG oslo.service.loopingcall [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.960685] env[61857]: DEBUG nova.compute.manager [-] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.960962] env[61857]: DEBUG nova.network.neutron [-] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 834.975743] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b6093530-32ef-488c-b4da-dfb82cfecca0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.976891] env[61857]: DEBUG oslo_vmware.api [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Task: {'id': task-951014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.349114} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.977802] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.978033] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 834.979094] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 834.979094] env[61857]: INFO nova.compute.manager [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Took 1.19 seconds to destroy the instance on the hypervisor. [ 834.979094] env[61857]: DEBUG oslo.service.loopingcall [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.979472] env[61857]: DEBUG nova.compute.manager [-] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.979526] env[61857]: DEBUG nova.network.neutron [-] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 834.982938] env[61857]: INFO nova.scheduler.client.report [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Deleted allocations for instance 2c187f64-8a36-4dfd-94e3-8ea944dbac24 [ 835.064479] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.064768] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.064841] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.065035] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.065207] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.065363] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.065569] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.065724] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.066465] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.066465] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.066465] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.072040] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8ea262b-a60c-4ea3-9bd6-4e7b7d919016 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.089423] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 835.089423] env[61857]: value = "task-951015" [ 835.089423] env[61857]: _type = "Task" [ 835.089423] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.097514] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951015, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.277333] env[61857]: DEBUG nova.compute.manager [req-33d6ba9a-388a-4a33-9087-e46833df9558 req-eccd9b58-4d1e-4b1d-8b59-ba7a7f3cb9ae service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Received event network-vif-deleted-3adcb286-015b-4b35-87f4-befacb5b4f3d {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 835.277333] env[61857]: INFO nova.compute.manager [req-33d6ba9a-388a-4a33-9087-e46833df9558 req-eccd9b58-4d1e-4b1d-8b59-ba7a7f3cb9ae service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Neutron deleted interface 3adcb286-015b-4b35-87f4-befacb5b4f3d; detaching it from the instance and deleting it from the info cache [ 835.277333] env[61857]: DEBUG nova.network.neutron [req-33d6ba9a-388a-4a33-9087-e46833df9558 req-eccd9b58-4d1e-4b1d-8b59-ba7a7f3cb9ae service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.292885] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951010, 'name': Rename_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.311753] env[61857]: DEBUG nova.compute.manager [req-d7a45a3e-dfeb-4cec-b3cb-2dcf80f59104 req-5fb716bb-a1b0-43c2-aa0e-0f27d8fa9226 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Received event network-vif-deleted-93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 835.312163] env[61857]: INFO nova.compute.manager [req-d7a45a3e-dfeb-4cec-b3cb-2dcf80f59104 req-5fb716bb-a1b0-43c2-aa0e-0f27d8fa9226 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Neutron deleted interface 93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb; detaching it from the instance and deleting it from the info cache [ 835.312608] env[61857]: DEBUG nova.network.neutron [req-d7a45a3e-dfeb-4cec-b3cb-2dcf80f59104 req-5fb716bb-a1b0-43c2-aa0e-0f27d8fa9226 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.317376] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Volume attach. Driver type: vmdk {{(pid=61857) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 835.318277] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214121', 'volume_id': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'name': 'volume-8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '36ca32ed-1ba2-40d0-93c4-086a26a3f9e5', 'attached_at': '', 'detached_at': '', 'volume_id': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'serial': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 835.321381] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8c24ef-8a84-49ae-8bb9-f3e714e5e38e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.341385] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7627fd10-81ee-461f-ad22-951702ad0546 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.362715] env[61857]: DEBUG nova.compute.manager [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.370485] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] volume-8e4ac73a-5a8b-4540-927e-a4d19d2affd9/volume-8e4ac73a-5a8b-4540-927e-a4d19d2affd9.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.371625] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcc1c8d-c0b2-4b6a-820b-a37f67261858 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.376798] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daf9111b-5542-492f-ab64-f5d54d99bfbf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.402867] env[61857]: DEBUG oslo_vmware.api [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 835.402867] env[61857]: value = "task-951016" [ 835.402867] env[61857]: _type = "Task" [ 835.402867] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.414108] env[61857]: DEBUG oslo_vmware.api [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951016, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.440283] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951007, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.494515] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6ca8db9-56bb-491e-942a-a4aacb9ff674 tempest-ServersV294TestFqdnHostnames-1396422356 tempest-ServersV294TestFqdnHostnames-1396422356-project-member] Lock "2c187f64-8a36-4dfd-94e3-8ea944dbac24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.277s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.599922] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951015, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.745953] env[61857]: DEBUG nova.network.neutron [-] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.779639] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e6ffb4d-cb2c-4d82-8972-5f39352e23a4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.791021] env[61857]: DEBUG nova.network.neutron [-] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.804348] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b30df8a-38c6-4467-82e7-3b7c96d13933 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.818798] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951010, 'name': Rename_Task, 'duration_secs': 1.239617} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.819512] env[61857]: INFO nova.compute.manager [-] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Took 0.86 seconds to deallocate network for instance. [ 835.820152] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 835.824015] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05a401c5-2b7b-4ffe-adb6-224d9b4f5f40 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.829020] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bb3e37a-0b6b-4497-bd14-ee795b038117 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.854503] env[61857]: DEBUG nova.compute.manager [req-33d6ba9a-388a-4a33-9087-e46833df9558 req-eccd9b58-4d1e-4b1d-8b59-ba7a7f3cb9ae service nova] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Detach interface failed, port_id=3adcb286-015b-4b35-87f4-befacb5b4f3d, reason: Instance a73f3163-b6fc-4764-b76a-451f823b5808 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 835.857900] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 835.857900] env[61857]: value = "task-951017" [ 835.857900] env[61857]: _type = "Task" [ 835.857900] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.861736] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5ce6f9-78d9-4f97-9eb2-4bcbc6a5badb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.887496] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951017, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.903070] env[61857]: DEBUG nova.compute.manager [req-d7a45a3e-dfeb-4cec-b3cb-2dcf80f59104 req-5fb716bb-a1b0-43c2-aa0e-0f27d8fa9226 service nova] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Detach interface failed, port_id=93a6a35f-9676-4b5f-bd15-8e1ce64b2ddb, reason: Instance bf2f46d4-5c6e-4493-b087-fe5a316b8c88 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 835.904311] env[61857]: INFO nova.compute.manager [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] instance snapshotting [ 835.910987] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04cf359-7674-4513-87bb-56e0ee725db7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.921558] env[61857]: DEBUG oslo_vmware.api [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951016, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.945456] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378c7f13-fac2-4653-968a-c4b1cb6aa780 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.954457] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951007, 'name': ReconfigVM_Task, 'duration_secs': 1.583552} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.957256] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Reconfigured VM instance instance-00000043 to attach disk [datastore2] ece27044-3123-4685-a4ab-30eeb10337d1/ece27044-3123-4685-a4ab-30eeb10337d1.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.961023] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa01fb93-da91-4470-a92c-a42d6b3f41f3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.969885] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 835.969885] env[61857]: value = "task-951018" [ 835.969885] env[61857]: _type = "Task" [ 835.969885] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.984097] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951018, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.100625] env[61857]: DEBUG nova.network.neutron [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Successfully updated port: 68fc598e-90c4-40b6-b89a-8e4fe65dd7dd {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.106233] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951015, 'name': ReconfigVM_Task, 'duration_secs': 0.514685} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.106985] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance '30d905ed-831d-44ae-807c-062de9a7e9fb' progress to 33 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 836.252074] env[61857]: INFO nova.compute.manager [-] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Took 1.27 seconds to deallocate network for instance. [ 836.333763] env[61857]: DEBUG oslo_concurrency.lockutils [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.386546] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951017, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.417724] env[61857]: DEBUG oslo_vmware.api [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951016, 'name': ReconfigVM_Task, 'duration_secs': 0.805933} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.418162] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Reconfigured VM instance instance-00000037 to attach disk [datastore2] volume-8e4ac73a-5a8b-4540-927e-a4d19d2affd9/volume-8e4ac73a-5a8b-4540-927e-a4d19d2affd9.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.423062] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67e47525-05ef-45a2-90f3-e5e1afa0664b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.442178] env[61857]: DEBUG oslo_vmware.api [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 836.442178] env[61857]: value = "task-951019" [ 836.442178] env[61857]: _type = "Task" [ 836.442178] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.453483] env[61857]: DEBUG oslo_vmware.api [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951019, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.463291] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 836.467170] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1f6b1323-1567-4aa1-aa53-f14951ed17b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.476378] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 836.476378] env[61857]: value = "task-951020" [ 836.476378] env[61857]: _type = "Task" [ 836.476378] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.485043] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951018, 'name': Rename_Task, 'duration_secs': 0.18139} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.487189] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.488234] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405c2b11-52f7-459f-b96e-2698eb3a0193 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.495414] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c791999-b4dc-423b-b251-c1fec8be6d27 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.497298] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951020, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.503770] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea16e63-a5a5-44b6-959b-06e805dcadae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.507358] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 836.507358] env[61857]: value = "task-951021" [ 836.507358] env[61857]: _type = "Task" [ 836.507358] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.540455] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744a5817-ba42-4675-8d9b-12730ee38e38 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.547599] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951021, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.556021] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fc7cb0-bde5-4337-8d0a-2535fce286fb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.567343] env[61857]: DEBUG nova.compute.provider_tree [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.615360] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.616411] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.617115] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.617115] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.617518] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.618679] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.618679] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.618679] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.618914] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.619821] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.620109] env[61857]: DEBUG nova.virt.hardware [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.625955] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Reconfiguring VM instance instance-00000039 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 836.626820] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64d6a7cb-b812-4911-b237-602a7dcec159 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.652355] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 836.652355] env[61857]: value = "task-951022" [ 836.652355] env[61857]: _type = "Task" [ 836.652355] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.662078] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951022, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.760308] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.889845] env[61857]: DEBUG oslo_vmware.api [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951017, 'name': PowerOnVM_Task, 'duration_secs': 0.655258} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.890262] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 836.894244] env[61857]: INFO nova.compute.manager [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Took 9.93 seconds to spawn the instance on the hypervisor. [ 836.894500] env[61857]: DEBUG nova.compute.manager [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.895477] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aaa8b25-94f3-4d9a-9c13-199a3f554129 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.954861] env[61857]: DEBUG oslo_vmware.api [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951019, 'name': ReconfigVM_Task, 'duration_secs': 0.171798} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.955372] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214121', 'volume_id': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'name': 'volume-8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '36ca32ed-1ba2-40d0-93c4-086a26a3f9e5', 'attached_at': '', 'detached_at': '', 'volume_id': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'serial': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 836.990538] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951020, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.017851] env[61857]: DEBUG oslo_vmware.api [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951021, 'name': PowerOnVM_Task, 'duration_secs': 0.47836} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.018160] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 837.018370] env[61857]: INFO nova.compute.manager [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Took 7.41 seconds to spawn the instance on the hypervisor. [ 837.018554] env[61857]: DEBUG nova.compute.manager [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.019369] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e0446e-00ca-46ec-a768-56fbb4c9c1e0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.070686] env[61857]: DEBUG nova.scheduler.client.report [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.166315] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951022, 'name': ReconfigVM_Task, 'duration_secs': 0.202295} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.166858] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Reconfigured VM instance instance-00000039 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 837.167829] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2951b637-f1ad-4ab5-b7fb-71130d1b35b2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.194530] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 30d905ed-831d-44ae-807c-062de9a7e9fb/30d905ed-831d-44ae-807c-062de9a7e9fb.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.195170] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5473a7a-25d6-4e52-949a-2a987045f380 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.219941] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 837.219941] env[61857]: value = "task-951023" [ 837.219941] env[61857]: _type = "Task" [ 837.219941] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.227844] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951023, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.420414] env[61857]: DEBUG nova.compute.manager [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Received event network-vif-plugged-68fc598e-90c4-40b6-b89a-8e4fe65dd7dd {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.420686] env[61857]: DEBUG oslo_concurrency.lockutils [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] Acquiring lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.420915] env[61857]: DEBUG oslo_concurrency.lockutils [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] Lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.421439] env[61857]: DEBUG oslo_concurrency.lockutils [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] Lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.421439] env[61857]: DEBUG nova.compute.manager [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] No waiting events found dispatching network-vif-plugged-68fc598e-90c4-40b6-b89a-8e4fe65dd7dd {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 837.421798] env[61857]: WARNING nova.compute.manager [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Received unexpected event network-vif-plugged-68fc598e-90c4-40b6-b89a-8e4fe65dd7dd for instance with vm_state building and task_state spawning. [ 837.421798] env[61857]: DEBUG nova.compute.manager [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Received event network-changed-68fc598e-90c4-40b6-b89a-8e4fe65dd7dd {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.421936] env[61857]: DEBUG nova.compute.manager [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Refreshing instance network info cache due to event network-changed-68fc598e-90c4-40b6-b89a-8e4fe65dd7dd. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 837.422030] env[61857]: DEBUG oslo_concurrency.lockutils [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] Acquiring lock "refresh_cache-ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.422193] env[61857]: DEBUG oslo_concurrency.lockutils [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] Acquired lock "refresh_cache-ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.422346] env[61857]: DEBUG nova.network.neutron [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Refreshing network info cache for port 68fc598e-90c4-40b6-b89a-8e4fe65dd7dd {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 837.423886] env[61857]: INFO nova.compute.manager [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Took 43.64 seconds to build instance. [ 837.492086] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951020, 'name': CreateSnapshot_Task, 'duration_secs': 0.836629} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.492505] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 837.496876] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943c20ce-df5f-4fc8-98c0-862114742c98 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.537343] env[61857]: INFO nova.compute.manager [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Took 42.39 seconds to build instance. [ 837.578760] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.578760] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 837.582656] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.744s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.584572] env[61857]: INFO nova.compute.claims [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.732163] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951023, 'name': ReconfigVM_Task, 'duration_secs': 0.428826} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.732490] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 30d905ed-831d-44ae-807c-062de9a7e9fb/30d905ed-831d-44ae-807c-062de9a7e9fb.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.732835] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance '30d905ed-831d-44ae-807c-062de9a7e9fb' progress to 50 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 837.929648] env[61857]: DEBUG oslo_concurrency.lockutils [None req-537f8f7b-4a95-4005-8207-f59744c0318b tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.137s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.015640] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 838.017052] env[61857]: DEBUG nova.network.neutron [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 838.019404] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-205041e7-a260-444f-a004-6c08f4e3c2d8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.028928] env[61857]: DEBUG nova.objects.instance [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lazy-loading 'flavor' on Instance uuid 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.031115] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 838.031115] env[61857]: value = "task-951024" [ 838.031115] env[61857]: _type = "Task" [ 838.031115] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.044599] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0b2dfef3-d004-4881-816f-77ab6b09d013 tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "ece27044-3123-4685-a4ab-30eeb10337d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.381s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.045410] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951024, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.096302] env[61857]: DEBUG nova.compute.utils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.097782] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 838.098191] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 838.208325] env[61857]: DEBUG nova.policy [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c021529e31ca4091a4a60001722a3262', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6a5f4ffeddf14cad8bd894501c9c624f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 838.221874] env[61857]: DEBUG nova.network.neutron [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.243069] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc671ffe-6df9-4258-a050-8c6168dad544 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.264127] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f2875c-ca2b-4a76-957d-69a47062a272 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.286580] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance '30d905ed-831d-44ae-807c-062de9a7e9fb' progress to 67 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 838.432289] env[61857]: DEBUG nova.compute.manager [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.545357] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68582499-d841-4231-97c6-cf46bd278081 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.844s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.551088] env[61857]: DEBUG nova.compute.manager [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.555886] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951024, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.601034] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 838.622823] env[61857]: DEBUG nova.network.neutron [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Successfully updated port: b19c85c5-d538-48ac-b311-493be32a9b2e {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.728474] env[61857]: DEBUG oslo_concurrency.lockutils [req-7a7e5f0d-6a89-4402-908b-154dea9201bc req-98a3557c-2239-4c93-b3c6-a1e73dc3368e service nova] Releasing lock "refresh_cache-ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.791853] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Successfully created port: 14eeb732-49e7-47b3-929e-85068b579c85 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 838.866308] env[61857]: DEBUG nova.network.neutron [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Port 9d8efcbb-1103-49f6-a52f-219e1682d275 binding to destination host cpu-1 is already ACTIVE {{(pid=61857) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 838.955510] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.048787] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951024, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.078583] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.125272] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "refresh_cache-ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.125412] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquired lock "refresh_cache-ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.125476] env[61857]: DEBUG nova.network.neutron [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 839.147073] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98a700f9-4e87-463b-8549-c7d05a8766f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.156789] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe6bf71-1bea-4921-a776-da01721be8da {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.194551] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc40a23-ee0e-4f20-84a3-a175fa2dd2f0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.204505] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17dc622d-8d38-4451-9f45-3980f91ba0bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.218896] env[61857]: DEBUG nova.compute.provider_tree [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.262335] env[61857]: INFO nova.compute.manager [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Rebuilding instance [ 839.316621] env[61857]: DEBUG nova.compute.manager [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.318103] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f45f1f-e4ba-417a-8bbe-d131848d01cd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.505154] env[61857]: DEBUG nova.compute.manager [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Received event network-changed-71f818ca-5a91-42ed-8825-5141bf75a30e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.505154] env[61857]: DEBUG nova.compute.manager [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Refreshing instance network info cache due to event network-changed-71f818ca-5a91-42ed-8825-5141bf75a30e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 839.505154] env[61857]: DEBUG oslo_concurrency.lockutils [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] Acquiring lock "refresh_cache-16b447b5-426a-4478-9d44-ae32b41dee50" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.505340] env[61857]: DEBUG oslo_concurrency.lockutils [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] Acquired lock "refresh_cache-16b447b5-426a-4478-9d44-ae32b41dee50" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.505525] env[61857]: DEBUG nova.network.neutron [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Refreshing network info cache for port 71f818ca-5a91-42ed-8825-5141bf75a30e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 839.546866] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951024, 'name': CloneVM_Task, 'duration_secs': 1.508086} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.547232] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Created linked-clone VM from snapshot [ 839.548070] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c78541-b600-4865-a50e-0b91acc86d3e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.557013] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Uploading image de6910ec-d5ea-4309-9229-4103bb53c96c {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 839.589499] env[61857]: DEBUG oslo_vmware.rw_handles [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 839.589499] env[61857]: value = "vm-214125" [ 839.589499] env[61857]: _type = "VirtualMachine" [ 839.589499] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 839.589499] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-70402820-47af-4f8a-9848-93cc719e7e37 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.595022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.595022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.597065] env[61857]: DEBUG oslo_vmware.rw_handles [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lease: (returnval){ [ 839.597065] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5203abde-260d-dea7-d310-9ae501d1a55d" [ 839.597065] env[61857]: _type = "HttpNfcLease" [ 839.597065] env[61857]: } obtained for exporting VM: (result){ [ 839.597065] env[61857]: value = "vm-214125" [ 839.597065] env[61857]: _type = "VirtualMachine" [ 839.597065] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 839.597426] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the lease: (returnval){ [ 839.597426] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5203abde-260d-dea7-d310-9ae501d1a55d" [ 839.597426] env[61857]: _type = "HttpNfcLease" [ 839.597426] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 839.606715] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 839.606715] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5203abde-260d-dea7-d310-9ae501d1a55d" [ 839.606715] env[61857]: _type = "HttpNfcLease" [ 839.606715] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 839.615030] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 839.638654] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.638932] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.639124] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.639548] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.639548] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.639881] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.639990] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.640166] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.640372] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.640568] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.640846] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.642080] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b5ed80-fc91-46ac-a893-ec52d3753252 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.653172] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63198fe-84eb-4b7a-97db-845b63f2aa6d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.669579] env[61857]: DEBUG nova.network.neutron [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 839.722623] env[61857]: DEBUG nova.scheduler.client.report [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.832441] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.832884] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63d1bec9-f62d-4e2e-a677-c19728b57386 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.842661] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 839.842661] env[61857]: value = "task-951026" [ 839.842661] env[61857]: _type = "Task" [ 839.842661] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.851277] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951026, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.890995] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "30d905ed-831d-44ae-807c-062de9a7e9fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.891316] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.891499] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.025704] env[61857]: DEBUG nova.network.neutron [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Updating instance_info_cache with network_info: [{"id": "68fc598e-90c4-40b6-b89a-8e4fe65dd7dd", "address": "fa:16:3e:b4:bc:bb", "network": {"id": "527d651e-2b58-4faa-a6e2-bd2c6e233095", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-471950059", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f520132b3ea465aa25f665b8a7750bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68fc598e-90", "ovs_interfaceid": "68fc598e-90c4-40b6-b89a-8e4fe65dd7dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b19c85c5-d538-48ac-b311-493be32a9b2e", "address": "fa:16:3e:c7:43:d2", "network": {"id": "cd57af49-404f-45d7-b917-eef44c42f3dc", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1203732448", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.182", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "5f520132b3ea465aa25f665b8a7750bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19c85c5-d5", "ovs_interfaceid": "b19c85c5-d538-48ac-b311-493be32a9b2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.101019] env[61857]: INFO nova.compute.manager [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Detaching volume 8e4ac73a-5a8b-4540-927e-a4d19d2affd9 [ 840.110837] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 840.110837] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5203abde-260d-dea7-d310-9ae501d1a55d" [ 840.110837] env[61857]: _type = "HttpNfcLease" [ 840.110837] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 840.113120] env[61857]: DEBUG oslo_vmware.rw_handles [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 840.113120] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5203abde-260d-dea7-d310-9ae501d1a55d" [ 840.113120] env[61857]: _type = "HttpNfcLease" [ 840.113120] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 840.113120] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74f5a62-73fa-45a1-a682-4631a11a03d7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.121459] env[61857]: DEBUG oslo_vmware.rw_handles [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc13b7-58a1-375b-b0c7-9051d5467866/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 840.121787] env[61857]: DEBUG oslo_vmware.rw_handles [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc13b7-58a1-375b-b0c7-9051d5467866/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 840.185780] env[61857]: INFO nova.virt.block_device [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Attempting to driver detach volume 8e4ac73a-5a8b-4540-927e-a4d19d2affd9 from mountpoint /dev/sdb [ 840.186127] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Volume detach. Driver type: vmdk {{(pid=61857) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 840.186370] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214121', 'volume_id': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'name': 'volume-8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '36ca32ed-1ba2-40d0-93c4-086a26a3f9e5', 'attached_at': '', 'detached_at': '', 'volume_id': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'serial': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 840.189440] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88b5ff6-8b72-4118-af41-69bd52cd5117 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.214675] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a494263-37db-41b7-9aa9-58aff4a1c084 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.222115] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7aadefe-b0f7-4cb0-adee-f22e504e1824 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.242510] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.660s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.243066] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.248062] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.142s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.248356] env[61857]: DEBUG nova.objects.instance [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lazy-loading 'resources' on Instance uuid d49008e6-7789-4785-8eea-9b3fa36c4ccb {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.253352] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a7a82c-6170-46b4-ac05-1a47e3c938c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.267843] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] The volume has not been displaced from its original location: [datastore2] volume-8e4ac73a-5a8b-4540-927e-a4d19d2affd9/volume-8e4ac73a-5a8b-4540-927e-a4d19d2affd9.vmdk. No consolidation needed. {{(pid=61857) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 840.273392] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Reconfiguring VM instance instance-00000037 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 840.276234] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c39bcd92-3302-4a71-990c-219e30e0e274 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.290224] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f4b3ffa3-c33b-4a7b-981f-a3df05fe7cfc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.297498] env[61857]: DEBUG oslo_vmware.api [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 840.297498] env[61857]: value = "task-951027" [ 840.297498] env[61857]: _type = "Task" [ 840.297498] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.313350] env[61857]: DEBUG oslo_vmware.api [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951027, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.352591] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951026, 'name': PowerOffVM_Task, 'duration_secs': 0.118891} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.352856] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 840.353081] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 840.353890] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93657dc9-023d-45ca-9fd6-61227f1aa419 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.362903] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 840.363177] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55376a97-0dcd-4d9b-bf8d-9ee31f72a3f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.386743] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 840.386938] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 840.387154] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Deleting the datastore file [datastore2] ece27044-3123-4685-a4ab-30eeb10337d1 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.387440] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7210ec6-ffa8-47e7-b0ab-e5b56e3911f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.394240] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 840.394240] env[61857]: value = "task-951029" [ 840.394240] env[61857]: _type = "Task" [ 840.394240] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.404078] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951029, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.509821] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Successfully updated port: 14eeb732-49e7-47b3-929e-85068b579c85 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.514972] env[61857]: DEBUG nova.network.neutron [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Updated VIF entry in instance network info cache for port 71f818ca-5a91-42ed-8825-5141bf75a30e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 840.514972] env[61857]: DEBUG nova.network.neutron [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Updating instance_info_cache with network_info: [{"id": "71f818ca-5a91-42ed-8825-5141bf75a30e", "address": "fa:16:3e:80:2f:1f", "network": {"id": "b4be0ad1-dff9-4093-8710-71550c9d3525", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1005551508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180bf8932fb845c68327ab92ead371d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2c68e7-b690-42e2-9491-c3f9357cc66a", "external-id": "nsx-vlan-transportzone-321", "segmentation_id": 321, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71f818ca-5a", "ovs_interfaceid": "71f818ca-5a91-42ed-8825-5141bf75a30e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.528131] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Releasing lock "refresh_cache-ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.529692] env[61857]: DEBUG nova.compute.manager [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Instance network_info: |[{"id": "68fc598e-90c4-40b6-b89a-8e4fe65dd7dd", "address": "fa:16:3e:b4:bc:bb", "network": {"id": "527d651e-2b58-4faa-a6e2-bd2c6e233095", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-471950059", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f520132b3ea465aa25f665b8a7750bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68fc598e-90", "ovs_interfaceid": "68fc598e-90c4-40b6-b89a-8e4fe65dd7dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b19c85c5-d538-48ac-b311-493be32a9b2e", "address": "fa:16:3e:c7:43:d2", "network": {"id": "cd57af49-404f-45d7-b917-eef44c42f3dc", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1203732448", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.182", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "5f520132b3ea465aa25f665b8a7750bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19c85c5-d5", "ovs_interfaceid": "b19c85c5-d538-48ac-b311-493be32a9b2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 840.531927] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:bc:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b107fab-ee71-47db-ad4d-3c6f05546843', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '68fc598e-90c4-40b6-b89a-8e4fe65dd7dd', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:43:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '572b7281-aad3-45fa-9cb2-fc1c70569948', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b19c85c5-d538-48ac-b311-493be32a9b2e', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.541977] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Creating folder: Project (5f520132b3ea465aa25f665b8a7750bd). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 840.546747] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-948be30f-131f-4e38-8b45-5ea6e99bb254 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.560053] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Created folder: Project (5f520132b3ea465aa25f665b8a7750bd) in parent group-v214027. [ 840.560393] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Creating folder: Instances. Parent ref: group-v214126. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 840.560656] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d26abd70-ccb1-49a4-b6e6-e883952244cc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.570944] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Created folder: Instances in parent group-v214126. [ 840.571603] env[61857]: DEBUG oslo.service.loopingcall [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.571603] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 840.571827] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9648890b-ded4-4d37-b864-eb32d917a30b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.598246] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.598246] env[61857]: value = "task-951032" [ 840.598246] env[61857]: _type = "Task" [ 840.598246] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.606467] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951032, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.756200] env[61857]: DEBUG nova.compute.utils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.764751] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.767450] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 840.814291] env[61857]: DEBUG oslo_vmware.api [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951027, 'name': ReconfigVM_Task, 'duration_secs': 0.288128} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.814810] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Reconfigured VM instance instance-00000037 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 840.825759] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56da91da-fd03-483d-b4d1-7419aaddedbb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.865947] env[61857]: DEBUG oslo_vmware.api [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 840.865947] env[61857]: value = "task-951033" [ 840.865947] env[61857]: _type = "Task" [ 840.865947] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.873222] env[61857]: DEBUG nova.policy [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c021529e31ca4091a4a60001722a3262', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6a5f4ffeddf14cad8bd894501c9c624f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.913443] env[61857]: DEBUG oslo_vmware.api [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951033, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.937597] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951029, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10401} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.940331] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.940331] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.944377] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 841.005443] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.005662] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.005867] env[61857]: DEBUG nova.network.neutron [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 841.018295] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "refresh_cache-d29f6e34-861d-4cbf-8cbd-ab193a55220b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.018725] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired lock "refresh_cache-d29f6e34-861d-4cbf-8cbd-ab193a55220b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.018725] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 841.020328] env[61857]: DEBUG oslo_concurrency.lockutils [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] Releasing lock "refresh_cache-16b447b5-426a-4478-9d44-ae32b41dee50" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.020783] env[61857]: DEBUG nova.compute.manager [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Received event network-vif-plugged-b19c85c5-d538-48ac-b311-493be32a9b2e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.021068] env[61857]: DEBUG oslo_concurrency.lockutils [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] Acquiring lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.021246] env[61857]: DEBUG oslo_concurrency.lockutils [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] Lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.021433] env[61857]: DEBUG oslo_concurrency.lockutils [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] Lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.022498] env[61857]: DEBUG nova.compute.manager [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] No waiting events found dispatching network-vif-plugged-b19c85c5-d538-48ac-b311-493be32a9b2e {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.022498] env[61857]: WARNING nova.compute.manager [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Received unexpected event network-vif-plugged-b19c85c5-d538-48ac-b311-493be32a9b2e for instance with vm_state building and task_state spawning. [ 841.022498] env[61857]: DEBUG nova.compute.manager [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Received event network-changed-b19c85c5-d538-48ac-b311-493be32a9b2e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.022498] env[61857]: DEBUG nova.compute.manager [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Refreshing instance network info cache due to event network-changed-b19c85c5-d538-48ac-b311-493be32a9b2e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 841.023183] env[61857]: DEBUG oslo_concurrency.lockutils [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] Acquiring lock "refresh_cache-ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.023380] env[61857]: DEBUG oslo_concurrency.lockutils [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] Acquired lock "refresh_cache-ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.023687] env[61857]: DEBUG nova.network.neutron [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Refreshing network info cache for port b19c85c5-d538-48ac-b311-493be32a9b2e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 841.114225] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951032, 'name': CreateVM_Task, 'duration_secs': 0.491738} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.114599] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 841.115453] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.115571] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.116101] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 841.119029] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-041ba14e-45f5-48e9-bc52-07e5a7fd9102 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.124757] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for the task: (returnval){ [ 841.124757] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525edbeb-3e12-c0b8-f2a1-f8275f1fbeca" [ 841.124757] env[61857]: _type = "Task" [ 841.124757] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.134938] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525edbeb-3e12-c0b8-f2a1-f8275f1fbeca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.264782] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.386136] env[61857]: DEBUG oslo_vmware.api [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951033, 'name': ReconfigVM_Task, 'duration_secs': 0.225566} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.390092] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214121', 'volume_id': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'name': 'volume-8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '36ca32ed-1ba2-40d0-93c4-086a26a3f9e5', 'attached_at': '', 'detached_at': '', 'volume_id': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9', 'serial': '8e4ac73a-5a8b-4540-927e-a4d19d2affd9'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 841.410051] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc0ae41-a346-42d3-b4b8-5f2d66540d47 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.419584] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d35609e-d84d-4746-b0e8-785ef0fb6bac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.459239] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c90bfe-8a77-44d2-a2ef-9d273c802c15 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.467707] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760f2aaa-358a-40ff-853c-c210b836fb44 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.483886] env[61857]: DEBUG nova.compute.provider_tree [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.523040] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Successfully created port: ec7e2b6d-16bc-4dc7-9355-f76d47ed088a {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.539261] env[61857]: DEBUG nova.compute.manager [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Received event network-vif-plugged-14eeb732-49e7-47b3-929e-85068b579c85 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.539585] env[61857]: DEBUG oslo_concurrency.lockutils [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] Acquiring lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.540225] env[61857]: DEBUG oslo_concurrency.lockutils [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] Lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.540225] env[61857]: DEBUG oslo_concurrency.lockutils [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] Lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.540376] env[61857]: DEBUG nova.compute.manager [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] No waiting events found dispatching network-vif-plugged-14eeb732-49e7-47b3-929e-85068b579c85 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.540546] env[61857]: WARNING nova.compute.manager [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Received unexpected event network-vif-plugged-14eeb732-49e7-47b3-929e-85068b579c85 for instance with vm_state building and task_state spawning. [ 841.540961] env[61857]: DEBUG nova.compute.manager [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Received event network-changed-14eeb732-49e7-47b3-929e-85068b579c85 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.541160] env[61857]: DEBUG nova.compute.manager [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Refreshing instance network info cache due to event network-changed-14eeb732-49e7-47b3-929e-85068b579c85. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 841.541416] env[61857]: DEBUG oslo_concurrency.lockutils [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] Acquiring lock "refresh_cache-d29f6e34-861d-4cbf-8cbd-ab193a55220b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.598748] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 841.640906] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525edbeb-3e12-c0b8-f2a1-f8275f1fbeca, 'name': SearchDatastore_Task, 'duration_secs': 0.010727} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.641432] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.641914] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.642428] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.642635] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.642905] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.643392] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-867b6d2b-3ab0-4c89-9780-7685b552d40e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.655820] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.656104] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 841.656930] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8fb583e-90f9-4588-8ba7-9c5349b60fce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.663950] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for the task: (returnval){ [ 841.663950] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52c82ed0-47a2-6568-da6b-e2ca0e910131" [ 841.663950] env[61857]: _type = "Task" [ 841.663950] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.673011] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c82ed0-47a2-6568-da6b-e2ca0e910131, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.958368] env[61857]: DEBUG nova.objects.instance [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lazy-loading 'flavor' on Instance uuid 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.988529] env[61857]: DEBUG nova.scheduler.client.report [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.993191] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Updating instance_info_cache with network_info: [{"id": "14eeb732-49e7-47b3-929e-85068b579c85", "address": "fa:16:3e:65:85:77", "network": {"id": "30eeaf02-1df3-4d59-83c2-af6bb436b328", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1408456233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5f4ffeddf14cad8bd894501c9c624f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14eeb732-49", "ovs_interfaceid": "14eeb732-49e7-47b3-929e-85068b579c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.045821] env[61857]: DEBUG nova.network.neutron [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Updated VIF entry in instance network info cache for port b19c85c5-d538-48ac-b311-493be32a9b2e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 842.046323] env[61857]: DEBUG nova.network.neutron [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Updating instance_info_cache with network_info: [{"id": "68fc598e-90c4-40b6-b89a-8e4fe65dd7dd", "address": "fa:16:3e:b4:bc:bb", "network": {"id": "527d651e-2b58-4faa-a6e2-bd2c6e233095", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-471950059", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f520132b3ea465aa25f665b8a7750bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68fc598e-90", "ovs_interfaceid": "68fc598e-90c4-40b6-b89a-8e4fe65dd7dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b19c85c5-d538-48ac-b311-493be32a9b2e", "address": "fa:16:3e:c7:43:d2", "network": {"id": "cd57af49-404f-45d7-b917-eef44c42f3dc", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1203732448", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.182", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "5f520132b3ea465aa25f665b8a7750bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19c85c5-d5", "ovs_interfaceid": "b19c85c5-d538-48ac-b311-493be32a9b2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.076223] env[61857]: DEBUG nova.network.neutron [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance_info_cache with network_info: [{"id": "9d8efcbb-1103-49f6-a52f-219e1682d275", "address": "fa:16:3e:54:79:1a", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d8efcbb-11", "ovs_interfaceid": "9d8efcbb-1103-49f6-a52f-219e1682d275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.177039] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c82ed0-47a2-6568-da6b-e2ca0e910131, 'name': SearchDatastore_Task, 'duration_secs': 0.009911} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.178039] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-811982f1-198e-4868-9a6a-aa8e63ea372c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.184708] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for the task: (returnval){ [ 842.184708] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52a90a2b-517b-205b-437f-fdbb42f14132" [ 842.184708] env[61857]: _type = "Task" [ 842.184708] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.194874] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a90a2b-517b-205b-437f-fdbb42f14132, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.283508] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.497662] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.249s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.500677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Releasing lock "refresh_cache-d29f6e34-861d-4cbf-8cbd-ab193a55220b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.501117] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Instance network_info: |[{"id": "14eeb732-49e7-47b3-929e-85068b579c85", "address": "fa:16:3e:65:85:77", "network": {"id": "30eeaf02-1df3-4d59-83c2-af6bb436b328", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1408456233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5f4ffeddf14cad8bd894501c9c624f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14eeb732-49", "ovs_interfaceid": "14eeb732-49e7-47b3-929e-85068b579c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 842.501277] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 30.498s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.504038] env[61857]: DEBUG oslo_concurrency.lockutils [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] Acquired lock "refresh_cache-d29f6e34-861d-4cbf-8cbd-ab193a55220b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.504038] env[61857]: DEBUG nova.network.neutron [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Refreshing network info cache for port 14eeb732-49e7-47b3-929e-85068b579c85 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 842.504641] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:85:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '14eeb732-49e7-47b3-929e-85068b579c85', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.513717] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Creating folder: Project (6a5f4ffeddf14cad8bd894501c9c624f). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 842.514960] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ac0bb1d0-756b-4faa-acc5-44141f2a5666 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.527499] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Created folder: Project (6a5f4ffeddf14cad8bd894501c9c624f) in parent group-v214027. [ 842.527815] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Creating folder: Instances. Parent ref: group-v214129. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 842.529469] env[61857]: INFO nova.scheduler.client.report [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Deleted allocations for instance d49008e6-7789-4785-8eea-9b3fa36c4ccb [ 842.530504] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-42f92342-feb3-434f-b5ea-f2700a9fcbd7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.544386] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Created folder: Instances in parent group-v214129. [ 842.544658] env[61857]: DEBUG oslo.service.loopingcall [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.544869] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 842.545363] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-12770554-627b-424b-b4db-a4b29df7b1a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.562340] env[61857]: DEBUG oslo_concurrency.lockutils [req-682b7055-43da-4b19-b12e-a8a800e99cd0 req-61da3e78-4db0-413c-8848-8a5d00a25cf0 service nova] Releasing lock "refresh_cache-ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.568561] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.568561] env[61857]: value = "task-951036" [ 842.568561] env[61857]: _type = "Task" [ 842.568561] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.579588] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.583396] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951036, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.699266] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a90a2b-517b-205b-437f-fdbb42f14132, 'name': SearchDatastore_Task, 'duration_secs': 0.010746} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.699608] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.699959] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] ad89cb5f-44d0-45c9-92a5-7f7ff3138e21/ad89cb5f-44d0-45c9-92a5-7f7ff3138e21.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 842.700336] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f3074f21-a0c1-48e9-a201-c0c3c17d1412 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.708034] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for the task: (returnval){ [ 842.708034] env[61857]: value = "task-951037" [ 842.708034] env[61857]: _type = "Task" [ 842.708034] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.718032] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951037, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.974965] env[61857]: DEBUG oslo_concurrency.lockutils [None req-64adff41-3b1c-40f4-87a8-2fa2624c018e tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.380s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.040989] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c16ba83e-dd26-4702-ba03-517936c58d42 tempest-ServerShowV247Test-430197403 tempest-ServerShowV247Test-430197403-project-member] Lock "d49008e6-7789-4785-8eea-9b3fa36c4ccb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.164s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.079993] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951036, 'name': CreateVM_Task, 'duration_secs': 0.455105} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.080279] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 843.081100] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.081293] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.081727] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.082102] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4513de94-61a0-44b9-bf0b-91f0d4e4d58b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.092964] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 843.092964] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523735f6-9542-47e0-f38e-53231df55f14" [ 843.092964] env[61857]: _type = "Task" [ 843.092964] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.104050] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523735f6-9542-47e0-f38e-53231df55f14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.111413] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffec5ed-793a-47f7-b5fb-7139168eba5c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.137890] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d8b9b9-fffc-416e-88c7-5fbcab56abbc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.144488] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance '30d905ed-831d-44ae-807c-062de9a7e9fb' progress to 83 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 843.219667] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951037, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466191} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.220538] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] ad89cb5f-44d0-45c9-92a5-7f7ff3138e21/ad89cb5f-44d0-45c9-92a5-7f7ff3138e21.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 843.220538] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.220751] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7473a947-3026-4562-b7bb-608e5e5d5bf7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.232031] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for the task: (returnval){ [ 843.232031] env[61857]: value = "task-951038" [ 843.232031] env[61857]: _type = "Task" [ 843.232031] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.239543] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951038, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.279694] env[61857]: DEBUG nova.network.neutron [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Updated VIF entry in instance network info cache for port 14eeb732-49e7-47b3-929e-85068b579c85. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 843.280387] env[61857]: DEBUG nova.network.neutron [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Updating instance_info_cache with network_info: [{"id": "14eeb732-49e7-47b3-929e-85068b579c85", "address": "fa:16:3e:65:85:77", "network": {"id": "30eeaf02-1df3-4d59-83c2-af6bb436b328", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1408456233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5f4ffeddf14cad8bd894501c9c624f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14eeb732-49", "ovs_interfaceid": "14eeb732-49e7-47b3-929e-85068b579c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.324871] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Successfully updated port: ec7e2b6d-16bc-4dc7-9355-f76d47ed088a {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.475190] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.475444] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.475622] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.475799] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.475950] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.476125] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.476348] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.476515] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.476689] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.476859] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.477052] env[61857]: DEBUG nova.virt.hardware [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.479421] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3a5ef8-352a-4b3b-9f62-441c532947e5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.489946] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.489946] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.489946] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.489946] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.489946] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.490292] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.490292] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.490425] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.490637] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.490840] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.491065] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.492377] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce2fd45-754c-4226-9d33-b347c6d5aab5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.498523] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d414f9-da75-48fd-ab26-2769178ee727 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.514224] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Instance VIF info [] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.520140] env[61857]: DEBUG oslo.service.loopingcall [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.522639] env[61857]: DEBUG oslo_vmware.rw_handles [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a91a3e-0e61-e244-9b8c-262e7e7a0092/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 843.524332] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 843.525016] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Applying migration context for instance 30d905ed-831d-44ae-807c-062de9a7e9fb as it has an incoming, in-progress migration f0272d66-8a8c-48a8-aad1-69a9053fe892. Migration status is post-migrating {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 843.527212] env[61857]: INFO nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating resource usage from migration f0272d66-8a8c-48a8-aad1-69a9053fe892 [ 843.529828] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9950c0d9-9cb2-4a90-9c87-cbed2f078e63 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.532839] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c82a4418-e211-4cd0-a7ba-cb0b814074cb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.551950] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbad101d-0687-4de7-949f-cef3416df555 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.558853] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e88aed62-6561-4e29-b041-46b19cc5ef63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.559103] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 18d9307d-6a20-47b4-b4e9-176f9a8c33cc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.559260] env[61857]: WARNING nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c4883f84-e62b-4b59-9484-270d82dc34e0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 843.559317] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 781a2790-e317-46fe-9be8-40c9e5f1f771 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.559452] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 71920a62-090d-4df1-937e-87df3b043e28 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.559606] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 28bf59ca-4ffe-4005-9a88-da0660ebb48a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.559728] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.559851] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e91e99b9-1cd3-4345-af09-f14af4df1214 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.559961] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 24d3d515-e03a-4b4e-bb8e-bc18537125ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.560082] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 42d38f5e-628a-4030-85e1-3ec0595cf3c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.560319] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 6c4c1276-ae80-4f37-9ef9-4872dd415d24 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.560474] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 96ff157c-4471-4a12-ad99-0aafd6c2dce5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.560641] env[61857]: WARNING nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 93e03e35-15a0-49e7-b1be-09178eabbfda is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 843.560767] env[61857]: WARNING nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance bf2f46d4-5c6e-4493-b087-fe5a316b8c88 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 843.560891] env[61857]: WARNING nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance a73f3163-b6fc-4764-b76a-451f823b5808 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 843.561013] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e5e64477-a1aa-4a4d-91a2-b17d912e09c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.561129] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 16b447b5-426a-4478-9d44-ae32b41dee50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.561243] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ece27044-3123-4685-a4ab-30eeb10337d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.561353] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ad89cb5f-44d0-45c9-92a5-7f7ff3138e21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.561460] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance d29f6e34-861d-4cbf-8cbd-ab193a55220b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.561568] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ddd4c436-405f-49f7-8c9b-de3b71725f63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.573607] env[61857]: DEBUG oslo_vmware.rw_handles [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a91a3e-0e61-e244-9b8c-262e7e7a0092/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 843.574094] env[61857]: ERROR oslo_vmware.rw_handles [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a91a3e-0e61-e244-9b8c-262e7e7a0092/disk-0.vmdk due to incomplete transfer. [ 843.574349] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.574349] env[61857]: value = "task-951039" [ 843.574349] env[61857]: _type = "Task" [ 843.574349] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.575297] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-44095cb9-72d0-49e7-9766-a1e8fa1f24fb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.586168] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951039, 'name': CreateVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.587227] env[61857]: DEBUG oslo_vmware.rw_handles [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a91a3e-0e61-e244-9b8c-262e7e7a0092/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 843.587433] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Uploaded image 31bb2f90-e5f9-4882-a312-c73ca6d91deb to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 843.589742] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 843.590093] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-447c50a9-5040-4e93-b4a2-c334c5e4a142 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.598551] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 843.598551] env[61857]: value = "task-951040" [ 843.598551] env[61857]: _type = "Task" [ 843.598551] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.605639] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523735f6-9542-47e0-f38e-53231df55f14, 'name': SearchDatastore_Task, 'duration_secs': 0.050196} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.606296] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.606545] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.606781] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.606930] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.607147] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.607411] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d6ca192b-2cba-4bf1-a65d-4f02f8705744 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.613673] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951040, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.622711] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.622929] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 843.623713] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ed55e03-b295-45a5-ba65-17c4b388cbc1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.629395] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 843.629395] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ea2846-c90b-2c06-f831-1445d618b35f" [ 843.629395] env[61857]: _type = "Task" [ 843.629395] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.637621] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ea2846-c90b-2c06-f831-1445d618b35f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.652488] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 843.652804] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7fdcb3a-32cf-40bd-9605-e64a1eed64b4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.659984] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 843.659984] env[61857]: value = "task-951041" [ 843.659984] env[61857]: _type = "Task" [ 843.659984] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.668582] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951041, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.740074] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951038, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071445} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.740470] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.742338] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108404fb-5838-46c6-8b18-268208d23e94 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.767188] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] ad89cb5f-44d0-45c9-92a5-7f7ff3138e21/ad89cb5f-44d0-45c9-92a5-7f7ff3138e21.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.767565] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c43b03c9-b921-40e0-96f3-11d79651173b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.793260] env[61857]: DEBUG oslo_concurrency.lockutils [req-95e2549e-1452-44d1-a8a4-38beb8e352a5 req-21c763ef-fac9-4138-b2c5-b66b938b570b service nova] Releasing lock "refresh_cache-d29f6e34-861d-4cbf-8cbd-ab193a55220b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.800293] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for the task: (returnval){ [ 843.800293] env[61857]: value = "task-951042" [ 843.800293] env[61857]: _type = "Task" [ 843.800293] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.809507] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951042, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.823609] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "refresh_cache-ddd4c436-405f-49f7-8c9b-de3b71725f63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.823930] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired lock "refresh_cache-ddd4c436-405f-49f7-8c9b-de3b71725f63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.824191] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.064897] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 7a9252da-2584-40ed-9d28-ca7341ed5165 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 844.090230] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951039, 'name': CreateVM_Task, 'duration_secs': 0.395583} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.090423] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 844.093062] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.093062] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.093062] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.093062] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0fde569-b1cf-45e0-9b6e-e2ed41875aaa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.098889] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 844.098889] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52c80f06-864a-32ac-380c-f6c65e130027" [ 844.098889] env[61857]: _type = "Task" [ 844.098889] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.110848] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951040, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.114542] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c80f06-864a-32ac-380c-f6c65e130027, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.140209] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ea2846-c90b-2c06-f831-1445d618b35f, 'name': SearchDatastore_Task, 'duration_secs': 0.031007} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.144021] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8a08157-b08a-44e0-b221-11060f6cc075 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.146776] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 844.146776] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52039008-f89e-3242-a6dc-923ed30a2544" [ 844.146776] env[61857]: _type = "Task" [ 844.146776] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.156552] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52039008-f89e-3242-a6dc-923ed30a2544, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.158459] env[61857]: DEBUG nova.compute.manager [req-3b5eb11c-e550-4053-ae38-b50b64805f5b req-8d88804b-527d-43cb-80c0-e7208121dd07 service nova] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Received event network-vif-plugged-ec7e2b6d-16bc-4dc7-9355-f76d47ed088a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.158909] env[61857]: DEBUG oslo_concurrency.lockutils [req-3b5eb11c-e550-4053-ae38-b50b64805f5b req-8d88804b-527d-43cb-80c0-e7208121dd07 service nova] Acquiring lock "ddd4c436-405f-49f7-8c9b-de3b71725f63-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.159244] env[61857]: DEBUG oslo_concurrency.lockutils [req-3b5eb11c-e550-4053-ae38-b50b64805f5b req-8d88804b-527d-43cb-80c0-e7208121dd07 service nova] Lock "ddd4c436-405f-49f7-8c9b-de3b71725f63-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.159475] env[61857]: DEBUG oslo_concurrency.lockutils [req-3b5eb11c-e550-4053-ae38-b50b64805f5b req-8d88804b-527d-43cb-80c0-e7208121dd07 service nova] Lock "ddd4c436-405f-49f7-8c9b-de3b71725f63-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.159666] env[61857]: DEBUG nova.compute.manager [req-3b5eb11c-e550-4053-ae38-b50b64805f5b req-8d88804b-527d-43cb-80c0-e7208121dd07 service nova] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] No waiting events found dispatching network-vif-plugged-ec7e2b6d-16bc-4dc7-9355-f76d47ed088a {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.159841] env[61857]: WARNING nova.compute.manager [req-3b5eb11c-e550-4053-ae38-b50b64805f5b req-8d88804b-527d-43cb-80c0-e7208121dd07 service nova] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Received unexpected event network-vif-plugged-ec7e2b6d-16bc-4dc7-9355-f76d47ed088a for instance with vm_state building and task_state spawning. [ 844.170357] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951041, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.311091] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951042, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.390060] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 844.571531] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c27f6f5b-6423-4eee-be53-dc48c93df893 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 844.584972] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Updating instance_info_cache with network_info: [{"id": "ec7e2b6d-16bc-4dc7-9355-f76d47ed088a", "address": "fa:16:3e:82:9d:ee", "network": {"id": "30eeaf02-1df3-4d59-83c2-af6bb436b328", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1408456233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5f4ffeddf14cad8bd894501c9c624f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec7e2b6d-16", "ovs_interfaceid": "ec7e2b6d-16bc-4dc7-9355-f76d47ed088a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.613362] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c80f06-864a-32ac-380c-f6c65e130027, 'name': SearchDatastore_Task, 'duration_secs': 0.013035} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.616462] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.616737] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.616958] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.620267] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951040, 'name': Destroy_Task, 'duration_secs': 0.543445} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.620267] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Destroyed the VM [ 844.620267] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 844.620267] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5e75c1c1-f05f-480f-84d1-64e3814eb188 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.625075] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 844.625075] env[61857]: value = "task-951043" [ 844.625075] env[61857]: _type = "Task" [ 844.625075] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.636339] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951043, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.658378] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52039008-f89e-3242-a6dc-923ed30a2544, 'name': SearchDatastore_Task, 'duration_secs': 0.028739} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.658731] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.659040] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] d29f6e34-861d-4cbf-8cbd-ab193a55220b/d29f6e34-861d-4cbf-8cbd-ab193a55220b.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 844.659405] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.659621] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.659853] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-782bdb8c-728b-4b79-95f3-132aa535cdba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.662694] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8da2faa-5a64-4c43-b2f2-803795cb6a73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.674598] env[61857]: DEBUG oslo_vmware.api [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951041, 'name': PowerOnVM_Task, 'duration_secs': 0.516577} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.676216] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 844.676435] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d95c7f46-863c-4942-8933-2c8a3c3462f2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance '30d905ed-831d-44ae-807c-062de9a7e9fb' progress to 100 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 844.680569] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 844.680569] env[61857]: value = "task-951044" [ 844.680569] env[61857]: _type = "Task" [ 844.680569] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.685034] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.685034] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 844.690296] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86bb62cc-7454-4859-885d-a00e0ba69718 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.699401] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 844.699401] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52a8c61c-8617-5a04-8d18-417b17086adf" [ 844.699401] env[61857]: _type = "Task" [ 844.699401] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.700024] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.710517] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a8c61c-8617-5a04-8d18-417b17086adf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.771138] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.771487] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.771777] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.772557] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.772557] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.774769] env[61857]: INFO nova.compute.manager [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Terminating instance [ 844.780026] env[61857]: DEBUG nova.compute.manager [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.780026] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 844.780026] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961d7e30-4653-426e-9f05-822a628e98ef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.787567] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 844.788033] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9193bae4-1aac-4528-84ec-1da7b6ed2bc9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.797164] env[61857]: DEBUG oslo_vmware.api [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 844.797164] env[61857]: value = "task-951045" [ 844.797164] env[61857]: _type = "Task" [ 844.797164] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.809900] env[61857]: DEBUG oslo_vmware.api [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951045, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.819337] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951042, 'name': ReconfigVM_Task, 'duration_secs': 0.789252} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.819337] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Reconfigured VM instance instance-00000044 to attach disk [datastore1] ad89cb5f-44d0-45c9-92a5-7f7ff3138e21/ad89cb5f-44d0-45c9-92a5-7f7ff3138e21.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.819337] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f3678da-ea8a-493b-a343-89ae67e49319 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.828239] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for the task: (returnval){ [ 844.828239] env[61857]: value = "task-951046" [ 844.828239] env[61857]: _type = "Task" [ 844.828239] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.836660] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951046, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.075714] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 3fb5d24b-1767-43d9-bed5-833121962adb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 845.088370] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Releasing lock "refresh_cache-ddd4c436-405f-49f7-8c9b-de3b71725f63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.088766] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Instance network_info: |[{"id": "ec7e2b6d-16bc-4dc7-9355-f76d47ed088a", "address": "fa:16:3e:82:9d:ee", "network": {"id": "30eeaf02-1df3-4d59-83c2-af6bb436b328", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1408456233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5f4ffeddf14cad8bd894501c9c624f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec7e2b6d-16", "ovs_interfaceid": "ec7e2b6d-16bc-4dc7-9355-f76d47ed088a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.089274] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:9d:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec7e2b6d-16bc-4dc7-9355-f76d47ed088a', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.098488] env[61857]: DEBUG oslo.service.loopingcall [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.102279] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 845.102672] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-24b8f1d6-fb36-457b-9f38-35ec52c9821e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.140384] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951043, 'name': RemoveSnapshot_Task} progress is 26%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.144735] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.144735] env[61857]: value = "task-951047" [ 845.144735] env[61857]: _type = "Task" [ 845.144735] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.154268] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951047, 'name': CreateVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.204249] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951044, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.216064] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a8c61c-8617-5a04-8d18-417b17086adf, 'name': SearchDatastore_Task, 'duration_secs': 0.014414} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.217979] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-981136dd-8916-4d4a-bf3a-3b49bab5fa45 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.223728] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 845.223728] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ca8563-50dc-11e7-9d79-5c0891413062" [ 845.223728] env[61857]: _type = "Task" [ 845.223728] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.233902] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ca8563-50dc-11e7-9d79-5c0891413062, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.258945] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquiring lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.259262] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.305807] env[61857]: DEBUG oslo_vmware.api [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951045, 'name': PowerOffVM_Task, 'duration_secs': 0.225995} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.307037] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 845.307037] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 845.307037] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd34a0cb-ac83-4101-a677-a6163e52d2db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.336414] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951046, 'name': Rename_Task, 'duration_secs': 0.211868} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.336792] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 845.336939] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78921fca-04b7-420d-b8af-30a76a56312d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.342441] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for the task: (returnval){ [ 845.342441] env[61857]: value = "task-951049" [ 845.342441] env[61857]: _type = "Task" [ 845.342441] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.350827] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951049, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.379230] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 845.379527] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 845.379686] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Deleting the datastore file [datastore2] 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.379991] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0c7aaae-7221-4688-b51e-f500746aa387 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.386990] env[61857]: DEBUG oslo_vmware.api [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 845.386990] env[61857]: value = "task-951050" [ 845.386990] env[61857]: _type = "Task" [ 845.386990] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.395997] env[61857]: DEBUG oslo_vmware.api [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951050, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.578918] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ddeed65b-9003-443f-8b2b-0756fbe7d234 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 845.579097] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Migration f0272d66-8a8c-48a8-aad1-69a9053fe892 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 845.579357] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 30d905ed-831d-44ae-807c-062de9a7e9fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 845.636885] env[61857]: DEBUG oslo_vmware.api [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951043, 'name': RemoveSnapshot_Task, 'duration_secs': 0.754836} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.638025] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 845.638025] env[61857]: INFO nova.compute.manager [None req-d82069cc-aae5-48ba-bb05-ab6f56e0c3cd tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Took 16.43 seconds to snapshot the instance on the hypervisor. [ 845.654711] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951047, 'name': CreateVM_Task, 'duration_secs': 0.387605} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.654895] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 845.655615] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.655799] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.656391] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.656391] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-095f3b75-f426-4505-8232-d9e395ff9949 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.661080] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 845.661080] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52340b4d-f158-c331-f94e-757c74c40701" [ 845.661080] env[61857]: _type = "Task" [ 845.661080] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.669314] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52340b4d-f158-c331-f94e-757c74c40701, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.695033] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951044, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531237} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.695365] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] d29f6e34-861d-4cbf-8cbd-ab193a55220b/d29f6e34-861d-4cbf-8cbd-ab193a55220b.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 845.695637] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.695934] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5411f59-e988-40e0-be67-608373f9d7c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.707680] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 845.707680] env[61857]: value = "task-951051" [ 845.707680] env[61857]: _type = "Task" [ 845.707680] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.716850] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951051, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.734770] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ca8563-50dc-11e7-9d79-5c0891413062, 'name': SearchDatastore_Task, 'duration_secs': 0.012966} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.735154] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.735560] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] ece27044-3123-4685-a4ab-30eeb10337d1/ece27044-3123-4685-a4ab-30eeb10337d1.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 845.735882] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3143d929-cf0b-4c91-a7af-95be3f651055 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.742423] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 845.742423] env[61857]: value = "task-951052" [ 845.742423] env[61857]: _type = "Task" [ 845.742423] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.753666] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951052, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.855579] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951049, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.897793] env[61857]: DEBUG oslo_vmware.api [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951050, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.342423} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.898107] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.898280] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 845.898564] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 845.898765] env[61857]: INFO nova.compute.manager [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Took 1.12 seconds to destroy the instance on the hypervisor. [ 845.899015] env[61857]: DEBUG oslo.service.loopingcall [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.899253] env[61857]: DEBUG nova.compute.manager [-] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.899327] env[61857]: DEBUG nova.network.neutron [-] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 846.084031] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ffa63de7-dd32-4908-92c5-755b3c0799ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 846.176029] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52340b4d-f158-c331-f94e-757c74c40701, 'name': SearchDatastore_Task, 'duration_secs': 0.016515} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.176029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.176029] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.176029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.176029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.176029] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.176029] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-079aa7c6-0441-415a-a61c-0ab4c6c3b6c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.193995] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.194304] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 846.195452] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf56b14d-0f06-4d6c-af08-d1d2da755203 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.202557] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 846.202557] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]527ba060-63d2-6fea-b0de-aeac2278805d" [ 846.202557] env[61857]: _type = "Task" [ 846.202557] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.231460] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]527ba060-63d2-6fea-b0de-aeac2278805d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.237731] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951051, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067252} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.238506] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.239854] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b65d5b1-fc6d-4538-9f13-43699ceecd0d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.272211] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] d29f6e34-861d-4cbf-8cbd-ab193a55220b/d29f6e34-861d-4cbf-8cbd-ab193a55220b.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.276493] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb396020-110b-4bce-9160-b3b6bae8f3dd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.292734] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951052, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.298737] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 846.298737] env[61857]: value = "task-951053" [ 846.298737] env[61857]: _type = "Task" [ 846.298737] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.307932] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951053, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.356237] env[61857]: DEBUG oslo_vmware.api [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951049, 'name': PowerOnVM_Task, 'duration_secs': 0.660948} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.356637] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 846.356854] env[61857]: INFO nova.compute.manager [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Took 11.62 seconds to spawn the instance on the hypervisor. [ 846.357064] env[61857]: DEBUG nova.compute.manager [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.358175] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec55a9d1-a87b-44e6-9c00-55e1e8ba1646 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.373386] env[61857]: DEBUG nova.compute.manager [req-4f46bcf7-bc36-4e6a-b904-2a935144c92b req-092f94cc-ae36-4bb9-b64d-37b092d501e1 service nova] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Received event network-changed-ec7e2b6d-16bc-4dc7-9355-f76d47ed088a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.373386] env[61857]: DEBUG nova.compute.manager [req-4f46bcf7-bc36-4e6a-b904-2a935144c92b req-092f94cc-ae36-4bb9-b64d-37b092d501e1 service nova] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Refreshing instance network info cache due to event network-changed-ec7e2b6d-16bc-4dc7-9355-f76d47ed088a. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 846.373386] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f46bcf7-bc36-4e6a-b904-2a935144c92b req-092f94cc-ae36-4bb9-b64d-37b092d501e1 service nova] Acquiring lock "refresh_cache-ddd4c436-405f-49f7-8c9b-de3b71725f63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.373386] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f46bcf7-bc36-4e6a-b904-2a935144c92b req-092f94cc-ae36-4bb9-b64d-37b092d501e1 service nova] Acquired lock "refresh_cache-ddd4c436-405f-49f7-8c9b-de3b71725f63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.373386] env[61857]: DEBUG nova.network.neutron [req-4f46bcf7-bc36-4e6a-b904-2a935144c92b req-092f94cc-ae36-4bb9-b64d-37b092d501e1 service nova] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Refreshing network info cache for port ec7e2b6d-16bc-4dc7-9355-f76d47ed088a {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 846.565441] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "30d905ed-831d-44ae-807c-062de9a7e9fb" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.565753] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.566202] env[61857]: DEBUG nova.compute.manager [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Going to confirm migration 1 {{(pid=61857) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 846.586056] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 6957dc98-8c60-4fdd-83dd-be2f13825c6d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 846.721021] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]527ba060-63d2-6fea-b0de-aeac2278805d, 'name': SearchDatastore_Task, 'duration_secs': 0.06435} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.721021] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ebe11a2-4b43-46a8-a601-67da31a6cc85 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.724711] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 846.724711] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad4c0d-8efa-a998-507b-00d8a3b9e171" [ 846.724711] env[61857]: _type = "Task" [ 846.724711] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.733573] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad4c0d-8efa-a998-507b-00d8a3b9e171, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.755679] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951052, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.750716} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.756158] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] ece27044-3123-4685-a4ab-30eeb10337d1/ece27044-3123-4685-a4ab-30eeb10337d1.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 846.756617] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.757019] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1feb85ea-24eb-4ca9-a371-be909bf58bd5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.765629] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 846.765629] env[61857]: value = "task-951054" [ 846.765629] env[61857]: _type = "Task" [ 846.765629] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.774545] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951054, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.812024] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951053, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.884539] env[61857]: INFO nova.compute.manager [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Took 47.57 seconds to build instance. [ 847.088877] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance d9d05629-79ef-45c6-ac54-b3bc18d306f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 847.089489] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 19 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 847.089763] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4224MB phys_disk=200GB used_disk=19GB total_vcpus=48 used_vcpus=19 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 847.140722] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.140722] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.140722] env[61857]: DEBUG nova.network.neutron [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.140722] env[61857]: DEBUG nova.objects.instance [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lazy-loading 'info_cache' on Instance uuid 30d905ed-831d-44ae-807c-062de9a7e9fb {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.156513] env[61857]: DEBUG nova.network.neutron [req-4f46bcf7-bc36-4e6a-b904-2a935144c92b req-092f94cc-ae36-4bb9-b64d-37b092d501e1 service nova] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Updated VIF entry in instance network info cache for port ec7e2b6d-16bc-4dc7-9355-f76d47ed088a. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 847.156513] env[61857]: DEBUG nova.network.neutron [req-4f46bcf7-bc36-4e6a-b904-2a935144c92b req-092f94cc-ae36-4bb9-b64d-37b092d501e1 service nova] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Updating instance_info_cache with network_info: [{"id": "ec7e2b6d-16bc-4dc7-9355-f76d47ed088a", "address": "fa:16:3e:82:9d:ee", "network": {"id": "30eeaf02-1df3-4d59-83c2-af6bb436b328", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1408456233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5f4ffeddf14cad8bd894501c9c624f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec7e2b6d-16", "ovs_interfaceid": "ec7e2b6d-16bc-4dc7-9355-f76d47ed088a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.240119] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad4c0d-8efa-a998-507b-00d8a3b9e171, 'name': SearchDatastore_Task, 'duration_secs': 0.029853} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.240119] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.240119] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] ddd4c436-405f-49f7-8c9b-de3b71725f63/ddd4c436-405f-49f7-8c9b-de3b71725f63.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 847.240479] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cf7f6fb4-694c-45b0-ab2c-39e1e6ab6d37 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.247320] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 847.247320] env[61857]: value = "task-951055" [ 847.247320] env[61857]: _type = "Task" [ 847.247320] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.259923] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951055, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.274957] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951054, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07769} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.276122] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.282775] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043b9c3e-4966-4fe0-b507-a3e3da294c1d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.303817] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] ece27044-3123-4685-a4ab-30eeb10337d1/ece27044-3123-4685-a4ab-30eeb10337d1.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.307663] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6fbc216b-4484-4859-abbe-2e1970daa6bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.323049] env[61857]: DEBUG nova.network.neutron [-] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.336025] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951053, 'name': ReconfigVM_Task, 'duration_secs': 0.563818} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.336504] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Reconfigured VM instance instance-00000045 to attach disk [datastore1] d29f6e34-861d-4cbf-8cbd-ab193a55220b/d29f6e34-861d-4cbf-8cbd-ab193a55220b.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.337328] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 847.337328] env[61857]: value = "task-951056" [ 847.337328] env[61857]: _type = "Task" [ 847.337328] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.340306] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3414d16b-21e9-4e20-b5a7-73c990bbf006 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.351641] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.355029] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 847.355029] env[61857]: value = "task-951057" [ 847.355029] env[61857]: _type = "Task" [ 847.355029] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.363501] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951057, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.386796] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7c098a8c-7075-428e-bd61-14c8a996745f tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.705s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.587697] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e251ed-1195-4ecf-9833-84773ff61655 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.596730] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587deef8-0c62-4e1f-a6ea-36e5972af9ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.635082] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d386e270-57e6-4ca6-8afe-6a0b34304c8d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.643202] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b97829f-c43d-4e64-9a62-34af5fc94eb4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.662024] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f46bcf7-bc36-4e6a-b904-2a935144c92b req-092f94cc-ae36-4bb9-b64d-37b092d501e1 service nova] Releasing lock "refresh_cache-ddd4c436-405f-49f7-8c9b-de3b71725f63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.662024] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.767631] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951055, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.829292] env[61857]: INFO nova.compute.manager [-] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Took 1.93 seconds to deallocate network for instance. [ 847.854493] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.864908] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951057, 'name': Rename_Task, 'duration_secs': 0.296823} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.865230] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 847.865534] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f31ca94a-13ad-4be1-a390-9b417337cad5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.875624] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 847.875624] env[61857]: value = "task-951058" [ 847.875624] env[61857]: _type = "Task" [ 847.875624] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.886189] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951058, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.893037] env[61857]: DEBUG nova.compute.manager [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 848.164680] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.205345] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.205544] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.205786] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.206010] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.206205] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.208530] env[61857]: INFO nova.compute.manager [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Terminating instance [ 848.211071] env[61857]: DEBUG nova.compute.manager [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 848.211319] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 848.212291] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9012a571-9d64-46c0-a449-28a1f1d50e45 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.220248] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 848.220461] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a0a1008-4f26-48e5-a0fd-61f7487c2c2d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.226150] env[61857]: DEBUG oslo_vmware.api [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for the task: (returnval){ [ 848.226150] env[61857]: value = "task-951059" [ 848.226150] env[61857]: _type = "Task" [ 848.226150] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.234554] env[61857]: DEBUG oslo_vmware.api [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951059, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.262082] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951055, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762415} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.262082] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] ddd4c436-405f-49f7-8c9b-de3b71725f63/ddd4c436-405f-49f7-8c9b-de3b71725f63.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 848.262307] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.262432] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ba5d225-897d-490e-a09e-fe25d518ade1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.268339] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 848.268339] env[61857]: value = "task-951060" [ 848.268339] env[61857]: _type = "Task" [ 848.268339] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.276583] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951060, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.293037] env[61857]: DEBUG nova.compute.manager [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.294083] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c7d3e7-be31-4665-9e03-8113b8a8a0fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.339215] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.353335] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951056, 'name': ReconfigVM_Task, 'duration_secs': 0.859411} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.353640] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Reconfigured VM instance instance-00000043 to attach disk [datastore1] ece27044-3123-4685-a4ab-30eeb10337d1/ece27044-3123-4685-a4ab-30eeb10337d1.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.354291] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5c0b7da-ba5d-496e-9112-35ead5248dce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.363151] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 848.363151] env[61857]: value = "task-951061" [ 848.363151] env[61857]: _type = "Task" [ 848.363151] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.371673] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951061, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.384991] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951058, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.403449] env[61857]: DEBUG nova.network.neutron [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance_info_cache with network_info: [{"id": "9d8efcbb-1103-49f6-a52f-219e1682d275", "address": "fa:16:3e:54:79:1a", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d8efcbb-11", "ovs_interfaceid": "9d8efcbb-1103-49f6-a52f-219e1682d275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.405445] env[61857]: DEBUG nova.compute.manager [req-71594924-f694-4dcb-91d5-cfb50c8a8b0a req-a457b561-20de-47df-92c8-a98f7dd2bc25 service nova] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Received event network-vif-deleted-f1227034-4a11-4433-ae59-a08bd1100d11 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 848.418109] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.669362] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 848.669646] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.168s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.670152] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.091s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.670561] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.672958] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.609s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.674555] env[61857]: INFO nova.compute.claims [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.695239] env[61857]: INFO nova.scheduler.client.report [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Deleted allocations for instance c4883f84-e62b-4b59-9484-270d82dc34e0 [ 848.736434] env[61857]: DEBUG oslo_vmware.api [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951059, 'name': PowerOffVM_Task, 'duration_secs': 0.274584} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.736748] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 848.736927] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 848.737245] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc49da72-37d2-4030-ba27-a612037d4696 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.779870] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951060, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112005} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.780118] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.781144] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73202535-6569-4fe0-ad93-9a58c60ab2f0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.804016] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] ddd4c436-405f-49f7-8c9b-de3b71725f63/ddd4c436-405f-49f7-8c9b-de3b71725f63.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.805067] env[61857]: INFO nova.compute.manager [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] instance snapshotting [ 848.806670] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ee053b0-5ed5-4792-a169-7f4a3a89bade {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.822149] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201b7ffa-9df4-40ad-a3b4-e72b714331bd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.844147] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bc820b-1ad9-4547-ab44-2a9e5b9ebc6a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.846966] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 848.846966] env[61857]: value = "task-951063" [ 848.846966] env[61857]: _type = "Task" [ 848.846966] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.859383] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951063, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.872027] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951061, 'name': Rename_Task, 'duration_secs': 0.209876} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.872331] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 848.872653] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b903afc-87df-4886-942e-e0680d805e4c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.878396] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 848.878396] env[61857]: value = "task-951064" [ 848.878396] env[61857]: _type = "Task" [ 848.878396] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.890599] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951058, 'name': PowerOnVM_Task, 'duration_secs': 0.723989} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.894022] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 848.894334] env[61857]: INFO nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Took 9.28 seconds to spawn the instance on the hypervisor. [ 848.894572] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.895291] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951064, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.896136] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bca049-c0d5-40e3-9e2b-1d4c371550b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.902214] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 848.902443] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 848.902645] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Deleting the datastore file [datastore1] ad89cb5f-44d0-45c9-92a5-7f7ff3138e21 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 848.904272] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5fdb295-a9d8-4643-b4b5-f44b4c527370 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.909166] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-30d905ed-831d-44ae-807c-062de9a7e9fb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.909417] env[61857]: DEBUG nova.objects.instance [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lazy-loading 'migration_context' on Instance uuid 30d905ed-831d-44ae-807c-062de9a7e9fb {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.916029] env[61857]: DEBUG oslo_vmware.api [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for the task: (returnval){ [ 848.916029] env[61857]: value = "task-951065" [ 848.916029] env[61857]: _type = "Task" [ 848.916029] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.925281] env[61857]: DEBUG oslo_vmware.api [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951065, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.205034] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4341d5d-bc7d-4e46-b98e-cc74e964656d tempest-ServersTestBootFromVolume-339672864 tempest-ServersTestBootFromVolume-339672864-project-member] Lock "c4883f84-e62b-4b59-9484-270d82dc34e0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.861s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.358708] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 849.358708] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951063, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.358789] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-12156864-38d9-4566-b859-854ad4407cc3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.365092] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 849.365092] env[61857]: value = "task-951066" [ 849.365092] env[61857]: _type = "Task" [ 849.365092] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.373773] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951066, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.392737] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951064, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.413685] env[61857]: DEBUG nova.objects.base [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Object Instance<30d905ed-831d-44ae-807c-062de9a7e9fb> lazy-loaded attributes: info_cache,migration_context {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 849.417931] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760aa67a-7ddd-4573-9ab3-23d5b3519f41 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.425888] env[61857]: INFO nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Took 44.60 seconds to build instance. [ 849.448740] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0646dc92-af1e-4e87-ba38-7c8dbba4d849 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.451601] env[61857]: DEBUG oslo_vmware.api [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Task: {'id': task-951065, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.355505} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.451861] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.452124] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 849.452317] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 849.452497] env[61857]: INFO nova.compute.manager [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Took 1.24 seconds to destroy the instance on the hypervisor. [ 849.453510] env[61857]: DEBUG oslo.service.loopingcall [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.453510] env[61857]: DEBUG nova.compute.manager [-] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.453623] env[61857]: DEBUG nova.network.neutron [-] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 849.457880] env[61857]: DEBUG oslo_vmware.api [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 849.457880] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520157d0-fb13-25ca-819c-1fd4737db0d6" [ 849.457880] env[61857]: _type = "Task" [ 849.457880] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.467211] env[61857]: DEBUG oslo_vmware.api [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]520157d0-fb13-25ca-819c-1fd4737db0d6, 'name': SearchDatastore_Task, 'duration_secs': 0.008907} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.467493] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.858375] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951063, 'name': ReconfigVM_Task, 'duration_secs': 0.592817} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.862476] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Reconfigured VM instance instance-00000046 to attach disk [datastore1] ddd4c436-405f-49f7-8c9b-de3b71725f63/ddd4c436-405f-49f7-8c9b-de3b71725f63.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.864514] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4de29a84-c9d7-450a-ba8d-4fc1be529d2c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.879036] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951066, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.880582] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 849.880582] env[61857]: value = "task-951067" [ 849.880582] env[61857]: _type = "Task" [ 849.880582] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.895419] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951067, 'name': Rename_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.899300] env[61857]: DEBUG oslo_vmware.api [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951064, 'name': PowerOnVM_Task, 'duration_secs': 0.769085} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.902467] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 849.902946] env[61857]: DEBUG nova.compute.manager [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.904226] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca247b21-73be-4cc6-84e0-11fb96becab0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.929179] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.826s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.163697] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bde1559-3a5a-4269-81c1-e5a7521814be {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.171244] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09539a3-a413-40d2-980f-94cb674d4ac8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.202605] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da30852-c73e-4ebf-adde-8ecdfa4e91d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.210102] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90155df2-05ae-4ea6-95ba-e3128965eec7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.224934] env[61857]: DEBUG nova.compute.provider_tree [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.381653] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951066, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.391481] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951067, 'name': Rename_Task, 'duration_secs': 0.19617} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.392664] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 850.392664] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b54d9455-2e3c-42f7-a663-9954a91d54bf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.399795] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 850.399795] env[61857]: value = "task-951068" [ 850.399795] env[61857]: _type = "Task" [ 850.399795] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.409869] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951068, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.424360] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.442914] env[61857]: DEBUG nova.network.neutron [-] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.469778] env[61857]: DEBUG nova.compute.manager [req-54693eeb-d16e-410a-9894-bdb340e0326b req-7f73cbf2-2e7e-48bc-ac99-35ff945677c3 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Received event network-vif-deleted-b19c85c5-d538-48ac-b311-493be32a9b2e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 850.469993] env[61857]: INFO nova.compute.manager [req-54693eeb-d16e-410a-9894-bdb340e0326b req-7f73cbf2-2e7e-48bc-ac99-35ff945677c3 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Neutron deleted interface b19c85c5-d538-48ac-b311-493be32a9b2e; detaching it from the instance and deleting it from the info cache [ 850.471328] env[61857]: DEBUG nova.network.neutron [req-54693eeb-d16e-410a-9894-bdb340e0326b req-7f73cbf2-2e7e-48bc-ac99-35ff945677c3 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Updating instance_info_cache with network_info: [{"id": "68fc598e-90c4-40b6-b89a-8e4fe65dd7dd", "address": "fa:16:3e:b4:bc:bb", "network": {"id": "527d651e-2b58-4faa-a6e2-bd2c6e233095", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-471950059", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f520132b3ea465aa25f665b8a7750bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68fc598e-90", "ovs_interfaceid": "68fc598e-90c4-40b6-b89a-8e4fe65dd7dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.728274] env[61857]: DEBUG nova.scheduler.client.report [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.881342] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951066, 'name': CreateSnapshot_Task, 'duration_secs': 1.095159} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.882613] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 850.883495] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a80b3d-dc45-48c9-afa4-c9c5b8f0e154 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.912293] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951068, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.945958] env[61857]: INFO nova.compute.manager [-] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Took 1.49 seconds to deallocate network for instance. [ 850.973462] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-99a07c5c-fac9-4aa5-ae3d-c01f9a9a7499 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.985913] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf658bcf-c08d-43bf-88fd-b044763f72c8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.023777] env[61857]: DEBUG nova.compute.manager [req-54693eeb-d16e-410a-9894-bdb340e0326b req-7f73cbf2-2e7e-48bc-ac99-35ff945677c3 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Detach interface failed, port_id=b19c85c5-d538-48ac-b311-493be32a9b2e, reason: Instance ad89cb5f-44d0-45c9-92a5-7f7ff3138e21 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 851.024187] env[61857]: DEBUG nova.compute.manager [req-54693eeb-d16e-410a-9894-bdb340e0326b req-7f73cbf2-2e7e-48bc-ac99-35ff945677c3 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Received event network-vif-deleted-68fc598e-90c4-40b6-b89a-8e4fe65dd7dd {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 851.024187] env[61857]: INFO nova.compute.manager [req-54693eeb-d16e-410a-9894-bdb340e0326b req-7f73cbf2-2e7e-48bc-ac99-35ff945677c3 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Neutron deleted interface 68fc598e-90c4-40b6-b89a-8e4fe65dd7dd; detaching it from the instance and deleting it from the info cache [ 851.024357] env[61857]: DEBUG nova.network.neutron [req-54693eeb-d16e-410a-9894-bdb340e0326b req-7f73cbf2-2e7e-48bc-ac99-35ff945677c3 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.235362] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.236117] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.239494] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.353s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.241816] env[61857]: INFO nova.compute.claims [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.248367] env[61857]: DEBUG oslo_vmware.rw_handles [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc13b7-58a1-375b-b0c7-9051d5467866/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 851.249629] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7aa9753-ffd4-46e1-98f1-286fd2d1466b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.257801] env[61857]: DEBUG oslo_vmware.rw_handles [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc13b7-58a1-375b-b0c7-9051d5467866/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 851.257801] env[61857]: ERROR oslo_vmware.rw_handles [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc13b7-58a1-375b-b0c7-9051d5467866/disk-0.vmdk due to incomplete transfer. [ 851.257947] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9cdc3502-853d-423b-9104-b5a985a3adca {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.274702] env[61857]: DEBUG oslo_vmware.rw_handles [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc13b7-58a1-375b-b0c7-9051d5467866/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 851.274937] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Uploaded image de6910ec-d5ea-4309-9229-4103bb53c96c to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 851.276939] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 851.279140] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-95075650-39af-46fd-9275-0dc38c8b09b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.291881] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 851.291881] env[61857]: value = "task-951069" [ 851.291881] env[61857]: _type = "Task" [ 851.291881] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.302405] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951069, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.328024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "ece27044-3123-4685-a4ab-30eeb10337d1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.328024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "ece27044-3123-4685-a4ab-30eeb10337d1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.328024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "ece27044-3123-4685-a4ab-30eeb10337d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.328024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "ece27044-3123-4685-a4ab-30eeb10337d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.328024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "ece27044-3123-4685-a4ab-30eeb10337d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.332041] env[61857]: INFO nova.compute.manager [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Terminating instance [ 851.334697] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "refresh_cache-ece27044-3123-4685-a4ab-30eeb10337d1" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.334954] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquired lock "refresh_cache-ece27044-3123-4685-a4ab-30eeb10337d1" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.335204] env[61857]: DEBUG nova.network.neutron [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 851.402899] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 851.403273] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-71ce00d0-f7e7-4694-8fc4-ddba8c1f0e38 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.415567] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951068, 'name': PowerOnVM_Task, 'duration_secs': 0.875585} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.416765] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 851.417140] env[61857]: INFO nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Took 9.13 seconds to spawn the instance on the hypervisor. [ 851.417230] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.417541] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 851.417541] env[61857]: value = "task-951070" [ 851.417541] env[61857]: _type = "Task" [ 851.417541] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.418303] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c78013-f05f-4dcc-9182-b06add5d76c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.432421] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951070, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.458031] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.529873] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f804117d-b20c-4785-9b99-998e639879d0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.540894] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e2ffb8-ffe7-450d-acc4-600ddc426a27 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.578527] env[61857]: DEBUG nova.compute.manager [req-54693eeb-d16e-410a-9894-bdb340e0326b req-7f73cbf2-2e7e-48bc-ac99-35ff945677c3 service nova] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Detach interface failed, port_id=68fc598e-90c4-40b6-b89a-8e4fe65dd7dd, reason: Instance ad89cb5f-44d0-45c9-92a5-7f7ff3138e21 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 851.746891] env[61857]: DEBUG nova.compute.utils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.750504] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 851.750737] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 851.802894] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951069, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.804485] env[61857]: DEBUG nova.policy [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c021529e31ca4091a4a60001722a3262', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6a5f4ffeddf14cad8bd894501c9c624f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 851.854085] env[61857]: DEBUG nova.network.neutron [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 851.903909] env[61857]: DEBUG nova.network.neutron [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.930735] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951070, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.942076] env[61857]: INFO nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Took 44.12 seconds to build instance. [ 852.149034] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Successfully created port: 3a8e7408-98e0-411a-8413-0968ac906574 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.254277] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.312023] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951069, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.410018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Releasing lock "refresh_cache-ece27044-3123-4685-a4ab-30eeb10337d1" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.410018] env[61857]: DEBUG nova.compute.manager [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 852.410018] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 852.410018] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e12a942-4132-4101-bdfd-37ccfe1cfca2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.419767] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 852.420066] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d63c475-bc7f-4c12-8571-1bc1deb06c48 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.431499] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951070, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.435601] env[61857]: DEBUG oslo_vmware.api [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 852.435601] env[61857]: value = "task-951071" [ 852.435601] env[61857]: _type = "Task" [ 852.435601] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.447476] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "ddd4c436-405f-49f7-8c9b-de3b71725f63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.311s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.447476] env[61857]: DEBUG oslo_vmware.api [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.747329] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3696ad8a-3fd7-4049-ae0d-c0fbe18aa477 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.755122] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d264ad5d-f0ab-4db8-9b9d-b1054ada62cd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.792123] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0831206e-a3f3-4cc5-a7d3-a604f5dfed8c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.807253] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74728fe8-55f4-4268-a9c9-5d491e4e3a3a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.811250] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951069, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.822342] env[61857]: DEBUG nova.compute.provider_tree [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.931790] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951070, 'name': CloneVM_Task} progress is 95%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.948667] env[61857]: DEBUG oslo_vmware.api [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951071, 'name': PowerOffVM_Task, 'duration_secs': 0.124916} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.949060] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 852.949253] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 852.949522] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4ad72a5c-0ee7-44c8-b49c-fa12410f45ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.984243] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 852.984492] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 852.984685] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Deleting the datastore file [datastore1] ece27044-3123-4685-a4ab-30eeb10337d1 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 852.984973] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc181f60-1117-493e-9e23-352b9e8b08ce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.990996] env[61857]: DEBUG oslo_vmware.api [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for the task: (returnval){ [ 852.990996] env[61857]: value = "task-951073" [ 852.990996] env[61857]: _type = "Task" [ 852.990996] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.999094] env[61857]: DEBUG oslo_vmware.api [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951073, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.293652] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.306651] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951069, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.321058] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.321377] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.321498] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.321705] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.321829] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.321983] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.322215] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.322409] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.322655] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.322822] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.323012] env[61857]: DEBUG nova.virt.hardware [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.323859] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04db4218-6f60-437c-89f3-437587bea566 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.327081] env[61857]: DEBUG nova.scheduler.client.report [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.336938] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6d74c0-80ca-4496-aa49-f48c4c89b692 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.381942] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "33cb5bbc-926d-42ee-b483-8d1e24707e40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.382207] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "33cb5bbc-926d-42ee-b483-8d1e24707e40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.432064] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951070, 'name': CloneVM_Task, 'duration_secs': 1.983271} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.432414] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Created linked-clone VM from snapshot [ 853.432967] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d249653f-d1ae-459c-ba0e-f1c1e11b2724 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.440513] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Uploading image 1bd5d088-3463-4ea9-8cc2-c14a6e3803f6 {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 853.461493] env[61857]: DEBUG oslo_vmware.rw_handles [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 853.461493] env[61857]: value = "vm-214135" [ 853.461493] env[61857]: _type = "VirtualMachine" [ 853.461493] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 853.461768] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-86b4f6ea-e0dd-46f9-b49a-5c2407297d5c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.469031] env[61857]: DEBUG oslo_vmware.rw_handles [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lease: (returnval){ [ 853.469031] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52474750-e40f-d9b2-8464-e9f6769939a7" [ 853.469031] env[61857]: _type = "HttpNfcLease" [ 853.469031] env[61857]: } obtained for exporting VM: (result){ [ 853.469031] env[61857]: value = "vm-214135" [ 853.469031] env[61857]: _type = "VirtualMachine" [ 853.469031] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 853.469288] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the lease: (returnval){ [ 853.469288] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52474750-e40f-d9b2-8464-e9f6769939a7" [ 853.469288] env[61857]: _type = "HttpNfcLease" [ 853.469288] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 853.475709] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 853.475709] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52474750-e40f-d9b2-8464-e9f6769939a7" [ 853.475709] env[61857]: _type = "HttpNfcLease" [ 853.475709] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 853.499787] env[61857]: DEBUG oslo_vmware.api [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Task: {'id': task-951073, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.285615} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.501042] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 853.501042] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 853.501042] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 853.501042] env[61857]: INFO nova.compute.manager [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Took 1.09 seconds to destroy the instance on the hypervisor. [ 853.501042] env[61857]: DEBUG oslo.service.loopingcall [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.501247] env[61857]: DEBUG nova.compute.manager [-] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 853.501277] env[61857]: DEBUG nova.network.neutron [-] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 853.519332] env[61857]: DEBUG nova.network.neutron [-] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.762409] env[61857]: DEBUG nova.compute.manager [req-c24888a2-95cb-4887-a4f3-3b6edb8f6688 req-8c111a0a-aa3a-4052-b98e-3d524c7622cf service nova] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Received event network-vif-plugged-3a8e7408-98e0-411a-8413-0968ac906574 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 853.762725] env[61857]: DEBUG oslo_concurrency.lockutils [req-c24888a2-95cb-4887-a4f3-3b6edb8f6688 req-8c111a0a-aa3a-4052-b98e-3d524c7622cf service nova] Acquiring lock "7a9252da-2584-40ed-9d28-ca7341ed5165-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.763033] env[61857]: DEBUG oslo_concurrency.lockutils [req-c24888a2-95cb-4887-a4f3-3b6edb8f6688 req-8c111a0a-aa3a-4052-b98e-3d524c7622cf service nova] Lock "7a9252da-2584-40ed-9d28-ca7341ed5165-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.763285] env[61857]: DEBUG oslo_concurrency.lockutils [req-c24888a2-95cb-4887-a4f3-3b6edb8f6688 req-8c111a0a-aa3a-4052-b98e-3d524c7622cf service nova] Lock "7a9252da-2584-40ed-9d28-ca7341ed5165-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.763534] env[61857]: DEBUG nova.compute.manager [req-c24888a2-95cb-4887-a4f3-3b6edb8f6688 req-8c111a0a-aa3a-4052-b98e-3d524c7622cf service nova] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] No waiting events found dispatching network-vif-plugged-3a8e7408-98e0-411a-8413-0968ac906574 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 853.763779] env[61857]: WARNING nova.compute.manager [req-c24888a2-95cb-4887-a4f3-3b6edb8f6688 req-8c111a0a-aa3a-4052-b98e-3d524c7622cf service nova] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Received unexpected event network-vif-plugged-3a8e7408-98e0-411a-8413-0968ac906574 for instance with vm_state building and task_state spawning. [ 853.807251] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951069, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.833649] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.834200] env[61857]: DEBUG nova.compute.manager [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 853.836679] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.682s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.838706] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.838946] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.643s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.840333] env[61857]: INFO nova.compute.claims [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.870422] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Successfully updated port: 3a8e7408-98e0-411a-8413-0968ac906574 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 853.881302] env[61857]: INFO nova.scheduler.client.report [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted allocations for instance 93e03e35-15a0-49e7-b1be-09178eabbfda [ 853.886457] env[61857]: DEBUG nova.compute.manager [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.977094] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 853.977094] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52474750-e40f-d9b2-8464-e9f6769939a7" [ 853.977094] env[61857]: _type = "HttpNfcLease" [ 853.977094] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 853.977451] env[61857]: DEBUG oslo_vmware.rw_handles [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 853.977451] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52474750-e40f-d9b2-8464-e9f6769939a7" [ 853.977451] env[61857]: _type = "HttpNfcLease" [ 853.977451] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 853.978284] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f63c96-0996-44c8-989e-3b8ec7e251d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.985920] env[61857]: DEBUG oslo_vmware.rw_handles [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283b47a-cacc-4a7b-c391-65bd1b0d2134/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 853.986160] env[61857]: DEBUG oslo_vmware.rw_handles [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283b47a-cacc-4a7b-c391-65bd1b0d2134/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 854.045190] env[61857]: DEBUG nova.network.neutron [-] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.092033] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6e437a49-fdf9-46ef-83cb-6c3eb8a5e100 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.307617] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951069, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.348030] env[61857]: DEBUG nova.compute.utils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.349123] env[61857]: DEBUG nova.compute.manager [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 854.349123] env[61857]: DEBUG nova.network.neutron [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 854.373801] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "refresh_cache-7a9252da-2584-40ed-9d28-ca7341ed5165" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.374392] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired lock "refresh_cache-7a9252da-2584-40ed-9d28-ca7341ed5165" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.374392] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.394302] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2d77e030-8b02-45b4-8be3-fc960063b99f tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "93e03e35-15a0-49e7-b1be-09178eabbfda" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.244s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.415179] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.424908] env[61857]: DEBUG nova.policy [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7aafb3356c5a48dcb36bf06d45851c3e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e3643abae2e40a99059603131e0b7b1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.548750] env[61857]: INFO nova.compute.manager [-] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Took 1.05 seconds to deallocate network for instance. [ 854.795865] env[61857]: DEBUG nova.network.neutron [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Successfully created port: d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.809364] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951069, 'name': Destroy_Task, 'duration_secs': 3.505196} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.809748] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Destroyed the VM [ 854.810061] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 854.810356] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3db47114-259b-4d56-8ded-bc3333ccfe9b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.817613] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 854.817613] env[61857]: value = "task-951075" [ 854.817613] env[61857]: _type = "Task" [ 854.817613] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.825815] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951075, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.853281] env[61857]: DEBUG nova.compute.manager [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 854.927650] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.055998] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.125880] env[61857]: DEBUG nova.network.neutron [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Updating instance_info_cache with network_info: [{"id": "3a8e7408-98e0-411a-8413-0968ac906574", "address": "fa:16:3e:ea:f6:d4", "network": {"id": "30eeaf02-1df3-4d59-83c2-af6bb436b328", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1408456233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5f4ffeddf14cad8bd894501c9c624f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8e7408-98", "ovs_interfaceid": "3a8e7408-98e0-411a-8413-0968ac906574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.329812] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951075, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.340061] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea35faf-3f9d-46be-8def-116fa0369a85 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.347536] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28fadf2-30e9-4330-8394-301f7534e662 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.386977] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712ae921-4bf0-47cb-805a-f3b32efe6d83 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.396631] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6aaef8-1261-406c-89f0-e7aa9ef334b0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.413867] env[61857]: DEBUG nova.compute.provider_tree [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.629068] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Releasing lock "refresh_cache-7a9252da-2584-40ed-9d28-ca7341ed5165" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.630214] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Instance network_info: |[{"id": "3a8e7408-98e0-411a-8413-0968ac906574", "address": "fa:16:3e:ea:f6:d4", "network": {"id": "30eeaf02-1df3-4d59-83c2-af6bb436b328", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1408456233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5f4ffeddf14cad8bd894501c9c624f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8e7408-98", "ovs_interfaceid": "3a8e7408-98e0-411a-8413-0968ac906574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.630367] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:f6:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a8e7408-98e0-411a-8413-0968ac906574', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.640519] env[61857]: DEBUG oslo.service.loopingcall [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.640903] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 855.641719] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-963d1ec9-4d43-477e-a4e7-a1ad90854b7d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.663804] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.663804] env[61857]: value = "task-951076" [ 855.663804] env[61857]: _type = "Task" [ 855.663804] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.672128] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951076, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.785388] env[61857]: DEBUG nova.compute.manager [req-aca36d05-2c85-49b2-ac29-1f334fd850dd req-4ddba99e-2ecb-4503-a0c0-1826d5c08cf8 service nova] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Received event network-changed-3a8e7408-98e0-411a-8413-0968ac906574 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 855.785802] env[61857]: DEBUG nova.compute.manager [req-aca36d05-2c85-49b2-ac29-1f334fd850dd req-4ddba99e-2ecb-4503-a0c0-1826d5c08cf8 service nova] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Refreshing instance network info cache due to event network-changed-3a8e7408-98e0-411a-8413-0968ac906574. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 855.786116] env[61857]: DEBUG oslo_concurrency.lockutils [req-aca36d05-2c85-49b2-ac29-1f334fd850dd req-4ddba99e-2ecb-4503-a0c0-1826d5c08cf8 service nova] Acquiring lock "refresh_cache-7a9252da-2584-40ed-9d28-ca7341ed5165" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.786420] env[61857]: DEBUG oslo_concurrency.lockutils [req-aca36d05-2c85-49b2-ac29-1f334fd850dd req-4ddba99e-2ecb-4503-a0c0-1826d5c08cf8 service nova] Acquired lock "refresh_cache-7a9252da-2584-40ed-9d28-ca7341ed5165" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.786597] env[61857]: DEBUG nova.network.neutron [req-aca36d05-2c85-49b2-ac29-1f334fd850dd req-4ddba99e-2ecb-4503-a0c0-1826d5c08cf8 service nova] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Refreshing network info cache for port 3a8e7408-98e0-411a-8413-0968ac906574 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 855.830072] env[61857]: DEBUG oslo_vmware.api [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951075, 'name': RemoveSnapshot_Task, 'duration_secs': 1.00077} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.831631] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 855.831631] env[61857]: INFO nova.compute.manager [None req-ad4664ef-b54d-4abb-aca2-01cfb8bc941e tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Took 19.92 seconds to snapshot the instance on the hypervisor. [ 855.888464] env[61857]: DEBUG nova.compute.manager [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 855.914849] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.915212] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.915423] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.915666] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.915891] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.916231] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.916477] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.916707] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.916904] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.917109] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.917365] env[61857]: DEBUG nova.virt.hardware [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.918127] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "d979a8a1-a5b5-428e-9b60-7d827337218f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.918472] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.920270] env[61857]: DEBUG nova.scheduler.client.report [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.923857] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f06faa-949f-4f42-a910-74c5822a720e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.934414] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8e308f-7ce0-402f-8812-81d7be3c5d36 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.176136] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951076, 'name': CreateVM_Task, 'duration_secs': 0.322464} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.176387] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 856.177118] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.177296] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.177685] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.177961] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2062052-c2a7-4137-9940-34a39531b4eb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.183347] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 856.183347] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52382c5f-baf4-06e4-e28e-28cb65cf6acc" [ 856.183347] env[61857]: _type = "Task" [ 856.183347] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.193734] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52382c5f-baf4-06e4-e28e-28cb65cf6acc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.391337] env[61857]: DEBUG nova.network.neutron [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Successfully updated port: d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.429015] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.429880] env[61857]: DEBUG nova.compute.manager [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.432625] env[61857]: DEBUG nova.compute.manager [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.441704] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.987s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.443279] env[61857]: INFO nova.compute.claims [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.619459] env[61857]: DEBUG nova.network.neutron [req-aca36d05-2c85-49b2-ac29-1f334fd850dd req-4ddba99e-2ecb-4503-a0c0-1826d5c08cf8 service nova] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Updated VIF entry in instance network info cache for port 3a8e7408-98e0-411a-8413-0968ac906574. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 856.619943] env[61857]: DEBUG nova.network.neutron [req-aca36d05-2c85-49b2-ac29-1f334fd850dd req-4ddba99e-2ecb-4503-a0c0-1826d5c08cf8 service nova] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Updating instance_info_cache with network_info: [{"id": "3a8e7408-98e0-411a-8413-0968ac906574", "address": "fa:16:3e:ea:f6:d4", "network": {"id": "30eeaf02-1df3-4d59-83c2-af6bb436b328", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1408456233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5f4ffeddf14cad8bd894501c9c624f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8e7408-98", "ovs_interfaceid": "3a8e7408-98e0-411a-8413-0968ac906574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.695416] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52382c5f-baf4-06e4-e28e-28cb65cf6acc, 'name': SearchDatastore_Task, 'duration_secs': 0.010257} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.695789] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.696140] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.696411] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.696578] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.696833] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.697528] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-308e9532-1b75-4215-9492-b26b0c00028b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.705838] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.706066] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 856.706937] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf4fdfa9-a9e5-4dc8-b51f-fa1b55e4346e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.713088] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 856.713088] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52109c9f-0b11-9862-add6-45fe58fba311" [ 856.713088] env[61857]: _type = "Task" [ 856.713088] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.721961] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52109c9f-0b11-9862-add6-45fe58fba311, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.894342] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.894513] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquired lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.894600] env[61857]: DEBUG nova.network.neutron [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.937737] env[61857]: DEBUG nova.compute.utils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.939299] env[61857]: DEBUG nova.compute.manager [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 856.939478] env[61857]: DEBUG nova.network.neutron [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 856.970340] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.983237] env[61857]: DEBUG nova.policy [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2ef7e845007475a8a19cf3f520c3f51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1acf476aabc4166bc8505a3442367c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.123263] env[61857]: DEBUG oslo_concurrency.lockutils [req-aca36d05-2c85-49b2-ac29-1f334fd850dd req-4ddba99e-2ecb-4503-a0c0-1826d5c08cf8 service nova] Releasing lock "refresh_cache-7a9252da-2584-40ed-9d28-ca7341ed5165" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.224367] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52109c9f-0b11-9862-add6-45fe58fba311, 'name': SearchDatastore_Task, 'duration_secs': 0.008831} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.225206] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a841aca0-6b17-405c-ad2c-33bb08b50b6e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.231057] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 857.231057] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52516db3-3860-0edb-19d8-cae9b861978c" [ 857.231057] env[61857]: _type = "Task" [ 857.231057] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.242162] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52516db3-3860-0edb-19d8-cae9b861978c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.259205] env[61857]: DEBUG nova.network.neutron [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Successfully created port: 472be223-da36-4834-9f18-05437505e7b8 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.445175] env[61857]: DEBUG nova.network.neutron [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 857.446010] env[61857]: DEBUG nova.compute.manager [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 857.677938] env[61857]: DEBUG nova.network.neutron [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Updating instance_info_cache with network_info: [{"id": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "address": "fa:16:3e:20:7f:b7", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1542c4d-16", "ovs_interfaceid": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.743644] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52516db3-3860-0edb-19d8-cae9b861978c, 'name': SearchDatastore_Task, 'duration_secs': 0.01019} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.746636] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.746906] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 7a9252da-2584-40ed-9d28-ca7341ed5165/7a9252da-2584-40ed-9d28-ca7341ed5165.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 857.747853] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8618f2f-afc6-40fc-b2f7-8f079b457517 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.756215] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 857.756215] env[61857]: value = "task-951077" [ 857.756215] env[61857]: _type = "Task" [ 857.756215] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.766590] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951077, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.936966] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682ff321-4759-43b2-8dcc-5ff4feef0835 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.947148] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0a0f79-1b9a-48ac-9c82-c5116e30b0e2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.987455] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad31b54-6382-4ce7-9be9-ea6644537b07 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.998306] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65460cb-aef6-440e-a876-54fa2acb2dbc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.014868] env[61857]: DEBUG nova.compute.provider_tree [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.067476] env[61857]: DEBUG nova.compute.manager [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Received event network-vif-plugged-d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 858.067780] env[61857]: DEBUG oslo_concurrency.lockutils [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] Acquiring lock "c27f6f5b-6423-4eee-be53-dc48c93df893-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.068020] env[61857]: DEBUG oslo_concurrency.lockutils [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] Lock "c27f6f5b-6423-4eee-be53-dc48c93df893-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.068330] env[61857]: DEBUG oslo_concurrency.lockutils [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] Lock "c27f6f5b-6423-4eee-be53-dc48c93df893-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.068557] env[61857]: DEBUG nova.compute.manager [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] No waiting events found dispatching network-vif-plugged-d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.068838] env[61857]: WARNING nova.compute.manager [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Received unexpected event network-vif-plugged-d1542c4d-1685-447b-a6c4-b56fad0ad098 for instance with vm_state building and task_state spawning. [ 858.069199] env[61857]: DEBUG nova.compute.manager [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Received event network-changed-d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 858.069426] env[61857]: DEBUG nova.compute.manager [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Refreshing instance network info cache due to event network-changed-d1542c4d-1685-447b-a6c4-b56fad0ad098. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 858.069631] env[61857]: DEBUG oslo_concurrency.lockutils [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] Acquiring lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.182085] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Releasing lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.182770] env[61857]: DEBUG nova.compute.manager [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Instance network_info: |[{"id": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "address": "fa:16:3e:20:7f:b7", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1542c4d-16", "ovs_interfaceid": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.183247] env[61857]: DEBUG oslo_concurrency.lockutils [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] Acquired lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.183499] env[61857]: DEBUG nova.network.neutron [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Refreshing network info cache for port d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 858.184948] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:7f:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4df917f7-847a-4c0e-b0e3-69a52e4a1554', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd1542c4d-1685-447b-a6c4-b56fad0ad098', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.194127] env[61857]: DEBUG oslo.service.loopingcall [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.195330] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 858.195615] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7a75727b-0cfd-41d6-b608-d94ef8d295a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.218874] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.218874] env[61857]: value = "task-951078" [ 858.218874] env[61857]: _type = "Task" [ 858.218874] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.229591] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951078, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.231382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "interface-96ff157c-4471-4a12-ad99-0aafd6c2dce5-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.231669] env[61857]: DEBUG oslo_concurrency.lockutils [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-96ff157c-4471-4a12-ad99-0aafd6c2dce5-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.232058] env[61857]: DEBUG nova.objects.instance [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'flavor' on Instance uuid 96ff157c-4471-4a12-ad99-0aafd6c2dce5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.266786] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951077, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500745} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.267083] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 7a9252da-2584-40ed-9d28-ca7341ed5165/7a9252da-2584-40ed-9d28-ca7341ed5165.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 858.267308] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 858.267573] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54d67d31-4f8a-443c-b9fb-ace7b44a994f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.275665] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 858.275665] env[61857]: value = "task-951079" [ 858.275665] env[61857]: _type = "Task" [ 858.275665] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.284611] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951079, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.458208] env[61857]: DEBUG nova.compute.manager [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.482745] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.483045] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.483217] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.483404] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.483555] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.483706] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.483919] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.484094] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.484268] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.484436] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.484608] env[61857]: DEBUG nova.virt.hardware [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.485529] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670057a1-c405-4e58-bbb2-6dfff5cf6eed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.494266] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7b1d22-1b00-468a-806e-2f0ac1f3496c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.518141] env[61857]: DEBUG nova.scheduler.client.report [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.730725] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951078, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.787260] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951079, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069868} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.787574] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 858.788669] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a969226e-7080-49c7-a2fb-d677407b2d85 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.828145] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 7a9252da-2584-40ed-9d28-ca7341ed5165/7a9252da-2584-40ed-9d28-ca7341ed5165.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 858.828519] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-553f3f55-ca2e-49a0-8e40-7df04c24a3a7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.853888] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 858.853888] env[61857]: value = "task-951080" [ 858.853888] env[61857]: _type = "Task" [ 858.853888] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.863029] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951080, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.864460] env[61857]: DEBUG nova.objects.instance [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'pci_requests' on Instance uuid 96ff157c-4471-4a12-ad99-0aafd6c2dce5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.962129] env[61857]: DEBUG nova.network.neutron [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Successfully updated port: 472be223-da36-4834-9f18-05437505e7b8 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.013321] env[61857]: DEBUG nova.network.neutron [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Updated VIF entry in instance network info cache for port d1542c4d-1685-447b-a6c4-b56fad0ad098. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 859.013751] env[61857]: DEBUG nova.network.neutron [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Updating instance_info_cache with network_info: [{"id": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "address": "fa:16:3e:20:7f:b7", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1542c4d-16", "ovs_interfaceid": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.023716] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.582s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.026287] env[61857]: DEBUG nova.compute.manager [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.028097] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.073s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.030930] env[61857]: INFO nova.compute.claims [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.096145] env[61857]: DEBUG nova.compute.manager [req-8af37c37-26e8-4b80-a7ad-4f2e66def6af req-db5cdba5-7fd8-447e-bd6b-7e39cc400e96 service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Received event network-vif-plugged-472be223-da36-4834-9f18-05437505e7b8 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 859.096614] env[61857]: DEBUG oslo_concurrency.lockutils [req-8af37c37-26e8-4b80-a7ad-4f2e66def6af req-db5cdba5-7fd8-447e-bd6b-7e39cc400e96 service nova] Acquiring lock "3fb5d24b-1767-43d9-bed5-833121962adb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.096927] env[61857]: DEBUG oslo_concurrency.lockutils [req-8af37c37-26e8-4b80-a7ad-4f2e66def6af req-db5cdba5-7fd8-447e-bd6b-7e39cc400e96 service nova] Lock "3fb5d24b-1767-43d9-bed5-833121962adb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.097362] env[61857]: DEBUG oslo_concurrency.lockutils [req-8af37c37-26e8-4b80-a7ad-4f2e66def6af req-db5cdba5-7fd8-447e-bd6b-7e39cc400e96 service nova] Lock "3fb5d24b-1767-43d9-bed5-833121962adb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.097605] env[61857]: DEBUG nova.compute.manager [req-8af37c37-26e8-4b80-a7ad-4f2e66def6af req-db5cdba5-7fd8-447e-bd6b-7e39cc400e96 service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] No waiting events found dispatching network-vif-plugged-472be223-da36-4834-9f18-05437505e7b8 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 859.097895] env[61857]: WARNING nova.compute.manager [req-8af37c37-26e8-4b80-a7ad-4f2e66def6af req-db5cdba5-7fd8-447e-bd6b-7e39cc400e96 service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Received unexpected event network-vif-plugged-472be223-da36-4834-9f18-05437505e7b8 for instance with vm_state building and task_state spawning. [ 859.229919] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951078, 'name': CreateVM_Task, 'duration_secs': 0.569732} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.229919] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 859.230361] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.230529] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.230837] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.231119] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7556cff0-067c-4327-bf0b-3390973c2e74 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.235812] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 859.235812] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523c3b1a-4ec0-db00-cc88-81d31667fff8" [ 859.235812] env[61857]: _type = "Task" [ 859.235812] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.243830] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523c3b1a-4ec0-db00-cc88-81d31667fff8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.364465] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951080, 'name': ReconfigVM_Task, 'duration_secs': 0.28484} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.364767] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 7a9252da-2584-40ed-9d28-ca7341ed5165/7a9252da-2584-40ed-9d28-ca7341ed5165.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.365410] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f6e632e-c808-4d26-9ab4-f80319531de6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.367219] env[61857]: DEBUG nova.objects.base [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Object Instance<96ff157c-4471-4a12-ad99-0aafd6c2dce5> lazy-loaded attributes: flavor,pci_requests {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 859.367417] env[61857]: DEBUG nova.network.neutron [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 859.373498] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 859.373498] env[61857]: value = "task-951081" [ 859.373498] env[61857]: _type = "Task" [ 859.373498] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.382087] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951081, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.465322] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "refresh_cache-3fb5d24b-1767-43d9-bed5-833121962adb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.465505] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "refresh_cache-3fb5d24b-1767-43d9-bed5-833121962adb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.465626] env[61857]: DEBUG nova.network.neutron [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 859.516881] env[61857]: DEBUG oslo_concurrency.lockutils [req-10513b7e-864f-427b-99cd-e1118e3e1846 req-3c72a01e-4a41-4350-99a1-2a57cc54620f service nova] Releasing lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.542031] env[61857]: DEBUG nova.compute.utils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.544931] env[61857]: DEBUG nova.compute.manager [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 859.546528] env[61857]: DEBUG nova.network.neutron [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 859.597150] env[61857]: DEBUG nova.policy [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd5c2aa088da5442b9332221da335f2d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4725d163dec74a53940dc24ae88927f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 859.657944] env[61857]: DEBUG nova.policy [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9056d344ff614950a60983450a2a40ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5653ed6ab0a948ef89bb03cf7426dd10', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 859.747139] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523c3b1a-4ec0-db00-cc88-81d31667fff8, 'name': SearchDatastore_Task, 'duration_secs': 0.010236} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.747555] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.747835] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.748301] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.748301] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.748601] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.748923] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da2e7997-de24-400b-b0ac-df102949abd2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.757617] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.757914] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 859.758750] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3bcf319-ff60-4496-a6a6-0ae44f7375ef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.764356] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 859.764356] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521f6cbc-3405-f74f-c760-31b729028214" [ 859.764356] env[61857]: _type = "Task" [ 859.764356] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.772880] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521f6cbc-3405-f74f-c760-31b729028214, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.869064] env[61857]: DEBUG nova.network.neutron [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Successfully created port: ff5c225b-875a-4492-b57d-a91e1063a36e {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.888028] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951081, 'name': Rename_Task, 'duration_secs': 0.129068} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.888420] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 859.888684] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91b18665-65e8-483d-aecb-63cf06339fe9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.895893] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 859.895893] env[61857]: value = "task-951082" [ 859.895893] env[61857]: _type = "Task" [ 859.895893] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.904265] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951082, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.011112] env[61857]: DEBUG nova.network.neutron [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 860.046769] env[61857]: DEBUG nova.compute.manager [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 860.060590] env[61857]: DEBUG nova.network.neutron [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Successfully created port: 78436af1-f52d-459b-8460-184ce130e879 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.282314] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521f6cbc-3405-f74f-c760-31b729028214, 'name': SearchDatastore_Task, 'duration_secs': 0.008953} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.284238] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f68df90b-5041-4085-860e-5d395a0ef7d0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.291138] env[61857]: DEBUG nova.network.neutron [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Updating instance_info_cache with network_info: [{"id": "472be223-da36-4834-9f18-05437505e7b8", "address": "fa:16:3e:9f:57:74", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap472be223-da", "ovs_interfaceid": "472be223-da36-4834-9f18-05437505e7b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.293665] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 860.293665] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]529ee8a5-78a0-6384-1caf-b65becaefe74" [ 860.293665] env[61857]: _type = "Task" [ 860.293665] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.318436] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529ee8a5-78a0-6384-1caf-b65becaefe74, 'name': SearchDatastore_Task, 'duration_secs': 0.015899} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.318895] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.319467] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] c27f6f5b-6423-4eee-be53-dc48c93df893/c27f6f5b-6423-4eee-be53-dc48c93df893.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 860.320800] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9d316df-a6ae-44bc-828f-42527218b787 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.330775] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 860.330775] env[61857]: value = "task-951083" [ 860.330775] env[61857]: _type = "Task" [ 860.330775] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.340513] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951083, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.407763] env[61857]: DEBUG oslo_vmware.api [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951082, 'name': PowerOnVM_Task, 'duration_secs': 0.475481} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.408128] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 860.408440] env[61857]: INFO nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Took 7.11 seconds to spawn the instance on the hypervisor. [ 860.408611] env[61857]: DEBUG nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.409548] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5852fac-3a22-45ac-b2b4-a95c4b1fbd9a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.541142] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8545be9f-6f72-4afd-a45e-f354903e2580 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.548891] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13a555c-9860-48ac-8564-5ca0b854253d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.584629] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383a10bc-1d8b-443a-a804-d737f749bbd1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.593498] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c6bd32-bd6d-485d-ac11-3dfb5969da91 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.608496] env[61857]: DEBUG nova.compute.provider_tree [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.796060] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "refresh_cache-3fb5d24b-1767-43d9-bed5-833121962adb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.796421] env[61857]: DEBUG nova.compute.manager [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Instance network_info: |[{"id": "472be223-da36-4834-9f18-05437505e7b8", "address": "fa:16:3e:9f:57:74", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap472be223-da", "ovs_interfaceid": "472be223-da36-4834-9f18-05437505e7b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 860.797044] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:57:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '472be223-da36-4834-9f18-05437505e7b8', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.805422] env[61857]: DEBUG oslo.service.loopingcall [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.805655] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 860.806370] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5d7cdf0-bddb-49f9-b953-2e4081748f3f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.827423] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 860.827423] env[61857]: value = "task-951084" [ 860.827423] env[61857]: _type = "Task" [ 860.827423] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.839661] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951084, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.843476] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951083, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.930481] env[61857]: INFO nova.compute.manager [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Took 40.90 seconds to build instance. [ 861.085598] env[61857]: DEBUG nova.compute.manager [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.112901] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.113241] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.113463] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.113732] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.113974] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.114241] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.114522] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.114767] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.115022] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.116336] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.116336] env[61857]: DEBUG nova.virt.hardware [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.117109] env[61857]: DEBUG nova.scheduler.client.report [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.121596] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6112a56-58b4-4305-8371-d642c3b46cd0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.135019] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145950c4-4127-4764-9239-23ea25031a3a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.341038] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951083, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.738005} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.344941] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] c27f6f5b-6423-4eee-be53-dc48c93df893/c27f6f5b-6423-4eee-be53-dc48c93df893.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 861.345222] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.345463] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951084, 'name': CreateVM_Task, 'duration_secs': 0.463519} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.345678] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f21d406d-d061-457d-9575-a7f405c7e919 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.348149] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 861.348324] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.348522] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.348846] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 861.349483] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ea37d18-f422-420a-9299-aa49b9ca7e2b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.354148] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 861.354148] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5246134e-dbb6-8676-fd15-40756baf3244" [ 861.354148] env[61857]: _type = "Task" [ 861.354148] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.355309] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 861.355309] env[61857]: value = "task-951085" [ 861.355309] env[61857]: _type = "Task" [ 861.355309] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.369059] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5246134e-dbb6-8676-fd15-40756baf3244, 'name': SearchDatastore_Task, 'duration_secs': 0.011625} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.373311] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.373635] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 861.373941] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.374158] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.374402] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.374749] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951085, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.375010] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-022e0a33-4de3-408b-b0eb-b1fc25dc39ed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.383783] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.384013] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 861.384892] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aff18229-1c1d-4c3f-94b5-2314e3090e3b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.391865] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 861.391865] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b4d216-2551-251e-7e76-5c5a38797b83" [ 861.391865] env[61857]: _type = "Task" [ 861.391865] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.399604] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b4d216-2551-251e-7e76-5c5a38797b83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.434106] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b7ec0cb3-b34c-4994-aa46-6cff9c12129c tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "7a9252da-2584-40ed-9d28-ca7341ed5165" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.232s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.475517] env[61857]: DEBUG nova.compute.manager [req-e09a3977-7496-4c17-9baa-bc85f2dfd6a1 req-4e3550e5-2e42-4df4-ac55-4cebe1fa2320 service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Received event network-changed-472be223-da36-4834-9f18-05437505e7b8 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 861.475733] env[61857]: DEBUG nova.compute.manager [req-e09a3977-7496-4c17-9baa-bc85f2dfd6a1 req-4e3550e5-2e42-4df4-ac55-4cebe1fa2320 service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Refreshing instance network info cache due to event network-changed-472be223-da36-4834-9f18-05437505e7b8. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 861.475960] env[61857]: DEBUG oslo_concurrency.lockutils [req-e09a3977-7496-4c17-9baa-bc85f2dfd6a1 req-4e3550e5-2e42-4df4-ac55-4cebe1fa2320 service nova] Acquiring lock "refresh_cache-3fb5d24b-1767-43d9-bed5-833121962adb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.476161] env[61857]: DEBUG oslo_concurrency.lockutils [req-e09a3977-7496-4c17-9baa-bc85f2dfd6a1 req-4e3550e5-2e42-4df4-ac55-4cebe1fa2320 service nova] Acquired lock "refresh_cache-3fb5d24b-1767-43d9-bed5-833121962adb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.476341] env[61857]: DEBUG nova.network.neutron [req-e09a3977-7496-4c17-9baa-bc85f2dfd6a1 req-4e3550e5-2e42-4df4-ac55-4cebe1fa2320 service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Refreshing network info cache for port 472be223-da36-4834-9f18-05437505e7b8 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 861.627257] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.599s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.627847] env[61857]: DEBUG nova.compute.manager [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 861.630512] env[61857]: DEBUG oslo_concurrency.lockutils [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.297s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.634066] env[61857]: DEBUG oslo_concurrency.lockutils [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.634066] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.873s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.634066] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.635104] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.680s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.636669] env[61857]: INFO nova.compute.claims [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.662364] env[61857]: INFO nova.scheduler.client.report [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Deleted allocations for instance bf2f46d4-5c6e-4493-b087-fe5a316b8c88 [ 861.664221] env[61857]: INFO nova.scheduler.client.report [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Deleted allocations for instance a73f3163-b6fc-4764-b76a-451f823b5808 [ 861.788807] env[61857]: DEBUG nova.network.neutron [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Successfully updated port: ff5c225b-875a-4492-b57d-a91e1063a36e {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.833417] env[61857]: DEBUG nova.network.neutron [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Successfully updated port: 78436af1-f52d-459b-8460-184ce130e879 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.870710] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951085, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069612} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.871083] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.871926] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9063db-19b7-4a4a-a4a7-841d4619fd6a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.895980] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] c27f6f5b-6423-4eee-be53-dc48c93df893/c27f6f5b-6423-4eee-be53-dc48c93df893.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.896440] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d9d1800-02ad-4267-b012-2351acdb609e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.922534] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b4d216-2551-251e-7e76-5c5a38797b83, 'name': SearchDatastore_Task, 'duration_secs': 0.008696} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.924627] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 861.924627] env[61857]: value = "task-951086" [ 861.924627] env[61857]: _type = "Task" [ 861.924627] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.924863] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df134aa1-4443-415e-8b2b-571d661761e7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.934751] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 861.934751] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]522b61fe-0357-0e82-025b-9c29ad72d7dd" [ 861.934751] env[61857]: _type = "Task" [ 861.934751] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.938630] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.947410] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522b61fe-0357-0e82-025b-9c29ad72d7dd, 'name': SearchDatastore_Task, 'duration_secs': 0.010187} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.947746] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.948039] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 3fb5d24b-1767-43d9-bed5-833121962adb/3fb5d24b-1767-43d9-bed5-833121962adb.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 861.948327] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b3176ba-02f0-4ae0-9425-528dc623ba4f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.955323] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 861.955323] env[61857]: value = "task-951087" [ 861.955323] env[61857]: _type = "Task" [ 861.955323] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.963621] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951087, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.142681] env[61857]: DEBUG nova.compute.utils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.149389] env[61857]: DEBUG nova.compute.manager [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.149389] env[61857]: DEBUG nova.network.neutron [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 862.176753] env[61857]: DEBUG oslo_concurrency.lockutils [None req-530af4a8-d0cc-4075-a2cc-9b164deebdb1 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "bf2f46d4-5c6e-4493-b087-fe5a316b8c88" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.461s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.178752] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959e4a96-ed33-4ed2-9f51-2d14c2c50af7 tempest-MultipleCreateTestJSON-714233784 tempest-MultipleCreateTestJSON-714233784-project-member] Lock "a73f3163-b6fc-4764-b76a-451f823b5808" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.404s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.201355] env[61857]: DEBUG nova.policy [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd71828dc4ff54dc7939df4e759285943', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b51cf69f1c947c89a3969daba09361b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.208184] env[61857]: DEBUG nova.compute.manager [req-846e85ea-edd0-4924-9768-3e61ff2e6791 req-c282e517-98d6-4c8a-9f4f-9b87dc6a5775 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Received event network-vif-plugged-78436af1-f52d-459b-8460-184ce130e879 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 862.208429] env[61857]: DEBUG oslo_concurrency.lockutils [req-846e85ea-edd0-4924-9768-3e61ff2e6791 req-c282e517-98d6-4c8a-9f4f-9b87dc6a5775 service nova] Acquiring lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.208706] env[61857]: DEBUG oslo_concurrency.lockutils [req-846e85ea-edd0-4924-9768-3e61ff2e6791 req-c282e517-98d6-4c8a-9f4f-9b87dc6a5775 service nova] Lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.208897] env[61857]: DEBUG oslo_concurrency.lockutils [req-846e85ea-edd0-4924-9768-3e61ff2e6791 req-c282e517-98d6-4c8a-9f4f-9b87dc6a5775 service nova] Lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.209163] env[61857]: DEBUG nova.compute.manager [req-846e85ea-edd0-4924-9768-3e61ff2e6791 req-c282e517-98d6-4c8a-9f4f-9b87dc6a5775 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] No waiting events found dispatching network-vif-plugged-78436af1-f52d-459b-8460-184ce130e879 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 862.209279] env[61857]: WARNING nova.compute.manager [req-846e85ea-edd0-4924-9768-3e61ff2e6791 req-c282e517-98d6-4c8a-9f4f-9b87dc6a5775 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Received unexpected event network-vif-plugged-78436af1-f52d-459b-8460-184ce130e879 for instance with vm_state active and task_state None. [ 862.283429] env[61857]: DEBUG nova.network.neutron [req-e09a3977-7496-4c17-9baa-bc85f2dfd6a1 req-4e3550e5-2e42-4df4-ac55-4cebe1fa2320 service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Updated VIF entry in instance network info cache for port 472be223-da36-4834-9f18-05437505e7b8. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 862.285025] env[61857]: DEBUG nova.network.neutron [req-e09a3977-7496-4c17-9baa-bc85f2dfd6a1 req-4e3550e5-2e42-4df4-ac55-4cebe1fa2320 service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Updating instance_info_cache with network_info: [{"id": "472be223-da36-4834-9f18-05437505e7b8", "address": "fa:16:3e:9f:57:74", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap472be223-da", "ovs_interfaceid": "472be223-da36-4834-9f18-05437505e7b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.291330] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "refresh_cache-ddeed65b-9003-443f-8b2b-0756fbe7d234" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.291581] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "refresh_cache-ddeed65b-9003-443f-8b2b-0756fbe7d234" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.291672] env[61857]: DEBUG nova.network.neutron [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 862.344572] env[61857]: DEBUG oslo_concurrency.lockutils [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.344962] env[61857]: DEBUG oslo_concurrency.lockutils [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.345167] env[61857]: DEBUG nova.network.neutron [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 862.440028] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951086, 'name': ReconfigVM_Task, 'duration_secs': 0.34736} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.440028] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Reconfigured VM instance instance-00000048 to attach disk [datastore2] c27f6f5b-6423-4eee-be53-dc48c93df893/c27f6f5b-6423-4eee-be53-dc48c93df893.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.440296] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69597a22-8943-4c9a-af6b-23669f111b03 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.447128] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 862.447128] env[61857]: value = "task-951088" [ 862.447128] env[61857]: _type = "Task" [ 862.447128] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.455853] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951088, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.464780] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951087, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463606} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.465069] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 3fb5d24b-1767-43d9-bed5-833121962adb/3fb5d24b-1767-43d9-bed5-833121962adb.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 862.465305] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.465560] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-829eef79-039c-42bf-a98d-be9d6d05ded0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.471546] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 862.471546] env[61857]: value = "task-951089" [ 862.471546] env[61857]: _type = "Task" [ 862.471546] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.479616] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951089, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.525866] env[61857]: DEBUG nova.network.neutron [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Successfully created port: 7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.650296] env[61857]: DEBUG nova.compute.manager [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 862.788967] env[61857]: DEBUG oslo_concurrency.lockutils [req-e09a3977-7496-4c17-9baa-bc85f2dfd6a1 req-4e3550e5-2e42-4df4-ac55-4cebe1fa2320 service nova] Releasing lock "refresh_cache-3fb5d24b-1767-43d9-bed5-833121962adb" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.841143] env[61857]: DEBUG nova.network.neutron [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 862.887768] env[61857]: WARNING nova.network.neutron [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] 30b7d27c-744c-46a6-8769-b1b1c95306e4 already exists in list: networks containing: ['30b7d27c-744c-46a6-8769-b1b1c95306e4']. ignoring it [ 862.967892] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951088, 'name': Rename_Task, 'duration_secs': 0.165517} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.968993] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 862.968993] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dc1917f-84c2-4f8f-b9ac-a97830a3a209 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.978789] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 862.978789] env[61857]: value = "task-951090" [ 862.978789] env[61857]: _type = "Task" [ 862.978789] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.988915] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951089, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082483} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.990439] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.990845] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20689bf-7836-415b-bd9b-b9fe8970c8c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.001873] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951090, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.023873] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.024210] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.024563] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.024824] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.025062] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.040829] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 3fb5d24b-1767-43d9-bed5-833121962adb/3fb5d24b-1767-43d9-bed5-833121962adb.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.040829] env[61857]: INFO nova.compute.manager [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Terminating instance [ 863.042962] env[61857]: DEBUG nova.compute.manager [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 863.044029] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 863.046585] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-675c9bce-1157-4295-8bc5-ce48a107514a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.063758] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c58bb4-cd03-4cac-84a4-ddca24115360 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.073608] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 863.075135] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-369273a4-69f5-4c92-a14f-958769017dd8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.077080] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 863.077080] env[61857]: value = "task-951091" [ 863.077080] env[61857]: _type = "Task" [ 863.077080] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.082329] env[61857]: DEBUG nova.network.neutron [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Updating instance_info_cache with network_info: [{"id": "ff5c225b-875a-4492-b57d-a91e1063a36e", "address": "fa:16:3e:04:8b:76", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff5c225b-87", "ovs_interfaceid": "ff5c225b-875a-4492-b57d-a91e1063a36e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.089526] env[61857]: DEBUG oslo_vmware.api [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 863.089526] env[61857]: value = "task-951092" [ 863.089526] env[61857]: _type = "Task" [ 863.089526] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.095524] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951091, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.100695] env[61857]: DEBUG oslo_vmware.api [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.147056] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d402a5f3-bf78-423c-8f62-034fc6ee267d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.162272] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a42b62-5e40-414c-811d-6b05b0430646 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.205021] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503ae28d-3046-40e6-bb87-50796913a132 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.211048] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e8dcad-4c80-431f-a28b-383f9a967d98 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.225885] env[61857]: DEBUG nova.compute.provider_tree [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.317339] env[61857]: DEBUG nova.network.neutron [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Updating instance_info_cache with network_info: [{"id": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "address": "fa:16:3e:88:8c:fd", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea2b08a8-d6", "ovs_interfaceid": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "78436af1-f52d-459b-8460-184ce130e879", "address": "fa:16:3e:53:2c:a8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78436af1-f5", "ovs_interfaceid": "78436af1-f52d-459b-8460-184ce130e879", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.464664] env[61857]: DEBUG oslo_vmware.rw_handles [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283b47a-cacc-4a7b-c391-65bd1b0d2134/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 863.465589] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46dea6d1-f620-4870-86cb-39e1ab9c094f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.471942] env[61857]: DEBUG oslo_vmware.rw_handles [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283b47a-cacc-4a7b-c391-65bd1b0d2134/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 863.472147] env[61857]: ERROR oslo_vmware.rw_handles [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283b47a-cacc-4a7b-c391-65bd1b0d2134/disk-0.vmdk due to incomplete transfer. [ 863.472785] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ee448392-611f-4f3e-a000-734195487623 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.479718] env[61857]: DEBUG oslo_vmware.rw_handles [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283b47a-cacc-4a7b-c391-65bd1b0d2134/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 863.479928] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Uploaded image 1bd5d088-3463-4ea9-8cc2-c14a6e3803f6 to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 863.482529] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 863.485595] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ea9d41a3-715a-4ab3-96ab-c63cd1bec5e9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.494263] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951090, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.494263] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 863.494263] env[61857]: value = "task-951093" [ 863.494263] env[61857]: _type = "Task" [ 863.494263] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.502674] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951093, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.591376] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "refresh_cache-ddeed65b-9003-443f-8b2b-0756fbe7d234" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.591783] env[61857]: DEBUG nova.compute.manager [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Instance network_info: |[{"id": "ff5c225b-875a-4492-b57d-a91e1063a36e", "address": "fa:16:3e:04:8b:76", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff5c225b-87", "ovs_interfaceid": "ff5c225b-875a-4492-b57d-a91e1063a36e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 863.592545] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951091, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.592964] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:8b:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff5c225b-875a-4492-b57d-a91e1063a36e', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.600611] env[61857]: DEBUG oslo.service.loopingcall [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.603748] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 863.604032] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e80ed97-4264-40b3-aba3-ff86c94efc5a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.623363] env[61857]: DEBUG oslo_vmware.api [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951092, 'name': PowerOffVM_Task, 'duration_secs': 0.223159} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.624709] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 863.624946] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 863.625224] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.625224] env[61857]: value = "task-951094" [ 863.625224] env[61857]: _type = "Task" [ 863.625224] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.625496] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e383dc6e-11d0-442f-9fe5-067023b48e02 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.634633] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951094, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.663197] env[61857]: DEBUG nova.compute.manager [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 863.686621] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.686893] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.687116] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.687360] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.687559] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.687758] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.688032] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.688256] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.688479] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.688692] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.688916] env[61857]: DEBUG nova.virt.hardware [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.690256] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36212e73-b22a-423d-a52b-3a66c79b032e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.698780] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c21c6c3-8668-49d2-aeae-d94ddbe8e2c2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.714749] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 863.715072] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 863.715318] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Deleting the datastore file [datastore1] d29f6e34-861d-4cbf-8cbd-ab193a55220b {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.715643] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78b5dfca-25c8-477d-a31b-55e0482e57f8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.721937] env[61857]: DEBUG oslo_vmware.api [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 863.721937] env[61857]: value = "task-951096" [ 863.721937] env[61857]: _type = "Task" [ 863.721937] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.732972] env[61857]: DEBUG nova.scheduler.client.report [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.737190] env[61857]: DEBUG oslo_vmware.api [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951096, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.820246] env[61857]: DEBUG oslo_concurrency.lockutils [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.821235] env[61857]: DEBUG oslo_concurrency.lockutils [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.821486] env[61857]: DEBUG oslo_concurrency.lockutils [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.822718] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf46eea-8265-4a35-88c4-4d506b34385b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.849398] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.849864] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.850207] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.850649] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.850862] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.851267] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.851587] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.852046] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.853557] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.853557] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.853557] env[61857]: DEBUG nova.virt.hardware [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.863426] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Reconfiguring VM to attach interface {{(pid=61857) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 863.863873] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-022c3d00-3b84-41e9-8d60-576408754a8d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.883019] env[61857]: DEBUG oslo_vmware.api [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 863.883019] env[61857]: value = "task-951097" [ 863.883019] env[61857]: _type = "Task" [ 863.883019] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.891632] env[61857]: DEBUG oslo_vmware.api [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951097, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.991070] env[61857]: DEBUG oslo_vmware.api [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951090, 'name': PowerOnVM_Task, 'duration_secs': 0.85636} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.991811] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 863.994022] env[61857]: INFO nova.compute.manager [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Took 8.10 seconds to spawn the instance on the hypervisor. [ 863.994022] env[61857]: DEBUG nova.compute.manager [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.994022] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e31611e-3704-44c6-a486-30c858c3fac7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.021065] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951093, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.074720] env[61857]: DEBUG nova.network.neutron [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Successfully updated port: 7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 864.090865] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951091, 'name': ReconfigVM_Task, 'duration_secs': 0.725981} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.091409] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 3fb5d24b-1767-43d9-bed5-833121962adb/3fb5d24b-1767-43d9-bed5-833121962adb.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.092335] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf4e4d36-bafc-484d-9ec3-a1d56f963a86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.102613] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 864.102613] env[61857]: value = "task-951098" [ 864.102613] env[61857]: _type = "Task" [ 864.102613] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.115628] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951098, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.124162] env[61857]: DEBUG nova.compute.manager [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Received event network-vif-plugged-ff5c225b-875a-4492-b57d-a91e1063a36e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 864.124162] env[61857]: DEBUG oslo_concurrency.lockutils [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] Acquiring lock "ddeed65b-9003-443f-8b2b-0756fbe7d234-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.124665] env[61857]: DEBUG oslo_concurrency.lockutils [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] Lock "ddeed65b-9003-443f-8b2b-0756fbe7d234-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.125073] env[61857]: DEBUG oslo_concurrency.lockutils [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] Lock "ddeed65b-9003-443f-8b2b-0756fbe7d234-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.125850] env[61857]: DEBUG nova.compute.manager [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] No waiting events found dispatching network-vif-plugged-ff5c225b-875a-4492-b57d-a91e1063a36e {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 864.129079] env[61857]: WARNING nova.compute.manager [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Received unexpected event network-vif-plugged-ff5c225b-875a-4492-b57d-a91e1063a36e for instance with vm_state building and task_state spawning. [ 864.129079] env[61857]: DEBUG nova.compute.manager [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Received event network-changed-ff5c225b-875a-4492-b57d-a91e1063a36e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 864.129079] env[61857]: DEBUG nova.compute.manager [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Refreshing instance network info cache due to event network-changed-ff5c225b-875a-4492-b57d-a91e1063a36e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 864.129079] env[61857]: DEBUG oslo_concurrency.lockutils [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] Acquiring lock "refresh_cache-ddeed65b-9003-443f-8b2b-0756fbe7d234" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.129079] env[61857]: DEBUG oslo_concurrency.lockutils [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] Acquired lock "refresh_cache-ddeed65b-9003-443f-8b2b-0756fbe7d234" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.129079] env[61857]: DEBUG nova.network.neutron [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Refreshing network info cache for port ff5c225b-875a-4492-b57d-a91e1063a36e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.144270] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951094, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.233544] env[61857]: DEBUG oslo_vmware.api [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951096, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.282787} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.233842] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.234144] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 864.234461] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 864.234600] env[61857]: INFO nova.compute.manager [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Took 1.19 seconds to destroy the instance on the hypervisor. [ 864.234852] env[61857]: DEBUG oslo.service.loopingcall [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.235088] env[61857]: DEBUG nova.compute.manager [-] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.235182] env[61857]: DEBUG nova.network.neutron [-] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 864.238880] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.604s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.239345] env[61857]: DEBUG nova.compute.manager [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 864.242186] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.164s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.243851] env[61857]: INFO nova.compute.claims [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.396149] env[61857]: DEBUG oslo_vmware.api [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.507204] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951093, 'name': Destroy_Task, 'duration_secs': 0.900409} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.507511] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Destroyed the VM [ 864.507756] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 864.508205] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4b6d9fb3-0c5c-444f-ae8d-7d75d4357e93 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.514497] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 864.514497] env[61857]: value = "task-951099" [ 864.514497] env[61857]: _type = "Task" [ 864.514497] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.530668] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951099, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.533108] env[61857]: INFO nova.compute.manager [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Took 38.67 seconds to build instance. [ 864.576719] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.576898] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquired lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.577042] env[61857]: DEBUG nova.network.neutron [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.616600] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951098, 'name': Rename_Task, 'duration_secs': 0.336721} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.616880] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 864.618331] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-771f5d08-e679-4d5c-9661-bcd56e498340 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.625980] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 864.625980] env[61857]: value = "task-951100" [ 864.625980] env[61857]: _type = "Task" [ 864.625980] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.647257] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951100, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.655526] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951094, 'name': CreateVM_Task, 'duration_secs': 0.729486} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.655721] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 864.662026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.662026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.662026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 864.662026] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61bf4636-0080-4696-b890-c345e179ca1c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.663578] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 864.663578] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]528459cd-39ee-5502-935c-6ec624ac75fc" [ 864.663578] env[61857]: _type = "Task" [ 864.663578] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.674918] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528459cd-39ee-5502-935c-6ec624ac75fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.748482] env[61857]: DEBUG nova.compute.utils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 864.753928] env[61857]: DEBUG nova.compute.manager [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 864.753928] env[61857]: DEBUG nova.network.neutron [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 864.761246] env[61857]: DEBUG nova.compute.manager [req-36ebd566-211c-42cd-beb9-458ade5baabf req-ab032427-7413-4d52-ab9b-b9feb2b9a4d3 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Received event network-changed-78436af1-f52d-459b-8460-184ce130e879 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 864.761655] env[61857]: DEBUG nova.compute.manager [req-36ebd566-211c-42cd-beb9-458ade5baabf req-ab032427-7413-4d52-ab9b-b9feb2b9a4d3 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Refreshing instance network info cache due to event network-changed-78436af1-f52d-459b-8460-184ce130e879. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 864.762082] env[61857]: DEBUG oslo_concurrency.lockutils [req-36ebd566-211c-42cd-beb9-458ade5baabf req-ab032427-7413-4d52-ab9b-b9feb2b9a4d3 service nova] Acquiring lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.762448] env[61857]: DEBUG oslo_concurrency.lockutils [req-36ebd566-211c-42cd-beb9-458ade5baabf req-ab032427-7413-4d52-ab9b-b9feb2b9a4d3 service nova] Acquired lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.762795] env[61857]: DEBUG nova.network.neutron [req-36ebd566-211c-42cd-beb9-458ade5baabf req-ab032427-7413-4d52-ab9b-b9feb2b9a4d3 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Refreshing network info cache for port 78436af1-f52d-459b-8460-184ce130e879 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.853618] env[61857]: DEBUG nova.policy [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ffa39f01b1a4131a78a8f9734eae9b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '192c67ad46484cf1b609342eb31962e6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.900745] env[61857]: DEBUG oslo_vmware.api [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.028722] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951099, 'name': RemoveSnapshot_Task} progress is 16%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.034893] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04419350-af65-42ab-833f-796a0db348ff tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "c27f6f5b-6423-4eee-be53-dc48c93df893" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.678s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.130878] env[61857]: DEBUG nova.network.neutron [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Updated VIF entry in instance network info cache for port ff5c225b-875a-4492-b57d-a91e1063a36e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 865.131448] env[61857]: DEBUG nova.network.neutron [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Updating instance_info_cache with network_info: [{"id": "ff5c225b-875a-4492-b57d-a91e1063a36e", "address": "fa:16:3e:04:8b:76", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff5c225b-87", "ovs_interfaceid": "ff5c225b-875a-4492-b57d-a91e1063a36e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.150821] env[61857]: DEBUG nova.network.neutron [-] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.152868] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951100, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.156985] env[61857]: DEBUG nova.network.neutron [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 865.177795] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528459cd-39ee-5502-935c-6ec624ac75fc, 'name': SearchDatastore_Task, 'duration_secs': 0.012896} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.178185] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.178466] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.179379] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.179558] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.179786] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.180516] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f86a66e0-8de3-4b6a-83fe-d0bccd1fbbc8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.191065] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.191364] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 865.192147] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-929f6dc4-2798-46ed-8a58-20d2b5ea214f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.201281] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 865.201281] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52c8a179-a6b9-0616-1e72-0b68cbc474a1" [ 865.201281] env[61857]: _type = "Task" [ 865.201281] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.212685] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c8a179-a6b9-0616-1e72-0b68cbc474a1, 'name': SearchDatastore_Task, 'duration_secs': 0.01018} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.213739] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6866925c-6dc5-474b-93af-e43c5d35da2c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.220466] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 865.220466] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5202cbfe-dcd6-250c-199a-98ada6ea9630" [ 865.220466] env[61857]: _type = "Task" [ 865.220466] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.229795] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5202cbfe-dcd6-250c-199a-98ada6ea9630, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.253359] env[61857]: DEBUG nova.compute.manager [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 865.401043] env[61857]: DEBUG nova.network.neutron [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Successfully created port: feebcdf7-ef87-4705-90cf-ebb4415a92fc {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.415292] env[61857]: DEBUG oslo_vmware.api [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951097, 'name': ReconfigVM_Task, 'duration_secs': 1.082899} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.416115] env[61857]: DEBUG oslo_concurrency.lockutils [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.416469] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Reconfigured VM to attach interface {{(pid=61857) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 865.493715] env[61857]: DEBUG nova.network.neutron [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Updating instance_info_cache with network_info: [{"id": "7ede4f79-7428-48b0-a811-e8c064d224f0", "address": "fa:16:3e:9e:cc:3a", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ede4f79-74", "ovs_interfaceid": "7ede4f79-7428-48b0-a811-e8c064d224f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.524721] env[61857]: DEBUG oslo_vmware.api [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951099, 'name': RemoveSnapshot_Task, 'duration_secs': 0.771569} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.525442] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 865.525610] env[61857]: INFO nova.compute.manager [None req-33140184-254d-4d31-ac97-aca6f0a690df tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Took 16.70 seconds to snapshot the instance on the hypervisor. [ 865.647402] env[61857]: DEBUG oslo_concurrency.lockutils [req-aaaa8ce1-4e56-46c6-9481-c3078f3a70a0 req-d8349bbd-6901-4670-98ee-37eac5b62185 service nova] Releasing lock "refresh_cache-ddeed65b-9003-443f-8b2b-0756fbe7d234" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.657925] env[61857]: INFO nova.compute.manager [-] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Took 1.42 seconds to deallocate network for instance. [ 865.657990] env[61857]: DEBUG oslo_vmware.api [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951100, 'name': PowerOnVM_Task, 'duration_secs': 0.824677} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.661180] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 865.661398] env[61857]: INFO nova.compute.manager [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Took 7.20 seconds to spawn the instance on the hypervisor. [ 865.661946] env[61857]: DEBUG nova.compute.manager [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.671743] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d497f4a-01c5-47b9-880f-9bd231a2cb12 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.678163] env[61857]: DEBUG nova.network.neutron [req-36ebd566-211c-42cd-beb9-458ade5baabf req-ab032427-7413-4d52-ab9b-b9feb2b9a4d3 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Updated VIF entry in instance network info cache for port 78436af1-f52d-459b-8460-184ce130e879. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 865.679022] env[61857]: DEBUG nova.network.neutron [req-36ebd566-211c-42cd-beb9-458ade5baabf req-ab032427-7413-4d52-ab9b-b9feb2b9a4d3 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Updating instance_info_cache with network_info: [{"id": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "address": "fa:16:3e:88:8c:fd", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea2b08a8-d6", "ovs_interfaceid": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "78436af1-f52d-459b-8460-184ce130e879", "address": "fa:16:3e:53:2c:a8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78436af1-f5", "ovs_interfaceid": "78436af1-f52d-459b-8460-184ce130e879", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.737803] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5202cbfe-dcd6-250c-199a-98ada6ea9630, 'name': SearchDatastore_Task, 'duration_secs': 0.011889} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.742028] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.742411] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] ddeed65b-9003-443f-8b2b-0756fbe7d234/ddeed65b-9003-443f-8b2b-0756fbe7d234.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 865.743066] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84a4440a-68b0-4535-9ff7-b1ecda8dbc3c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.753044] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 865.753044] env[61857]: value = "task-951101" [ 865.753044] env[61857]: _type = "Task" [ 865.753044] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.775740] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951101, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.877949] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0604dcfc-4643-4591-a832-f51a0fcd026f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.887159] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7da83bd-cae4-4faa-a7a8-a905896bb8a0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.922888] env[61857]: DEBUG oslo_concurrency.lockutils [None req-845c06cf-4fce-4a9e-a3ce-5dc06deb1a62 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-96ff157c-4471-4a12-ad99-0aafd6c2dce5-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.691s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.925180] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41afa649-0c63-4e35-81fd-29a841474db3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.934836] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d043c797-224c-412d-9590-50ae6a0c3413 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.950359] env[61857]: DEBUG nova.compute.provider_tree [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.996247] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Releasing lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.996663] env[61857]: DEBUG nova.compute.manager [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Instance network_info: |[{"id": "7ede4f79-7428-48b0-a811-e8c064d224f0", "address": "fa:16:3e:9e:cc:3a", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ede4f79-74", "ovs_interfaceid": "7ede4f79-7428-48b0-a811-e8c064d224f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 865.997145] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:cc:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afd3feb3-ffcc-4499-a2c2-eb6a48aefde9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7ede4f79-7428-48b0-a811-e8c064d224f0', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.005714] env[61857]: DEBUG oslo.service.loopingcall [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.006465] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 866.006465] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f533f08-c240-47b8-a833-328a0901f0cf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.030072] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.030072] env[61857]: value = "task-951102" [ 866.030072] env[61857]: _type = "Task" [ 866.030072] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.043068] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951102, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.178483] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.183390] env[61857]: DEBUG oslo_concurrency.lockutils [req-36ebd566-211c-42cd-beb9-458ade5baabf req-ab032427-7413-4d52-ab9b-b9feb2b9a4d3 service nova] Releasing lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.202735] env[61857]: INFO nova.compute.manager [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Took 38.03 seconds to build instance. [ 866.265329] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951101, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.268686] env[61857]: DEBUG nova.compute.manager [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 866.298104] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.298375] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.298541] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.298731] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.298924] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.299093] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.299337] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.299479] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.299665] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.299894] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.300045] env[61857]: DEBUG nova.virt.hardware [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.301083] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8db0de-bf1d-467f-993e-28d27d14d542 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.310570] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc414f1b-67c0-41db-86d3-374a62e93385 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.454071] env[61857]: DEBUG nova.scheduler.client.report [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.543251] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951102, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.705755] env[61857]: DEBUG oslo_concurrency.lockutils [None req-16b1f5d9-185e-48f1-aac1-83e11986fbd7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "3fb5d24b-1767-43d9-bed5-833121962adb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.253s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.769246] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951101, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.827142} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.769246] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] ddeed65b-9003-443f-8b2b-0756fbe7d234/ddeed65b-9003-443f-8b2b-0756fbe7d234.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 866.769246] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.769246] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9da386c-b49c-4483-a0ea-b2a9de09df4d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.776791] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 866.776791] env[61857]: value = "task-951103" [ 866.776791] env[61857]: _type = "Task" [ 866.776791] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.788939] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951103, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.963066] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.719s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.963066] env[61857]: DEBUG nova.compute.manager [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.965533] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.626s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.966399] env[61857]: DEBUG nova.objects.instance [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lazy-loading 'resources' on Instance uuid 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.023108] env[61857]: DEBUG nova.compute.manager [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Received event network-vif-plugged-7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 867.023108] env[61857]: DEBUG oslo_concurrency.lockutils [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] Acquiring lock "ffa63de7-dd32-4908-92c5-755b3c0799ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.027473] env[61857]: DEBUG oslo_concurrency.lockutils [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.027473] env[61857]: DEBUG oslo_concurrency.lockutils [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.004s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.027473] env[61857]: DEBUG nova.compute.manager [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] No waiting events found dispatching network-vif-plugged-7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 867.027473] env[61857]: WARNING nova.compute.manager [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Received unexpected event network-vif-plugged-7ede4f79-7428-48b0-a811-e8c064d224f0 for instance with vm_state building and task_state spawning. [ 867.027473] env[61857]: DEBUG nova.compute.manager [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Received event network-changed-7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 867.027815] env[61857]: DEBUG nova.compute.manager [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Refreshing instance network info cache due to event network-changed-7ede4f79-7428-48b0-a811-e8c064d224f0. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 867.027931] env[61857]: DEBUG oslo_concurrency.lockutils [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] Acquiring lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.028011] env[61857]: DEBUG oslo_concurrency.lockutils [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] Acquired lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.028206] env[61857]: DEBUG nova.network.neutron [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Refreshing network info cache for port 7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 867.045450] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951102, 'name': CreateVM_Task, 'duration_secs': 0.585727} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.045726] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 867.046415] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.047426] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.047857] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.051220] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe0ec55c-fec1-4439-94a0-e5d7285d96de {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.059195] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 867.059195] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52d41e1d-1ebf-3502-9650-033b01a80c04" [ 867.059195] env[61857]: _type = "Task" [ 867.059195] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.068598] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52d41e1d-1ebf-3502-9650-033b01a80c04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.268103] env[61857]: DEBUG nova.network.neutron [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Successfully updated port: feebcdf7-ef87-4705-90cf-ebb4415a92fc {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.292244] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951103, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082323} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.292689] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.293722] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4159c2-9d04-4937-931e-4a7cd653b0b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.322678] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] ddeed65b-9003-443f-8b2b-0756fbe7d234/ddeed65b-9003-443f-8b2b-0756fbe7d234.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.323055] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c5ea670-d086-4442-a5fc-e399841bdd25 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.344472] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 867.344472] env[61857]: value = "task-951104" [ 867.344472] env[61857]: _type = "Task" [ 867.344472] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.352914] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951104, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.448456] env[61857]: DEBUG nova.compute.manager [req-9ba979e4-32e4-4b79-9ab8-283d22259446 req-ccb34d8e-e5f2-4e18-84cb-b8e4ef839717 service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Received event network-changed-d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 867.448649] env[61857]: DEBUG nova.compute.manager [req-9ba979e4-32e4-4b79-9ab8-283d22259446 req-ccb34d8e-e5f2-4e18-84cb-b8e4ef839717 service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Refreshing instance network info cache due to event network-changed-d1542c4d-1685-447b-a6c4-b56fad0ad098. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 867.448862] env[61857]: DEBUG oslo_concurrency.lockutils [req-9ba979e4-32e4-4b79-9ab8-283d22259446 req-ccb34d8e-e5f2-4e18-84cb-b8e4ef839717 service nova] Acquiring lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.449192] env[61857]: DEBUG oslo_concurrency.lockutils [req-9ba979e4-32e4-4b79-9ab8-283d22259446 req-ccb34d8e-e5f2-4e18-84cb-b8e4ef839717 service nova] Acquired lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.451462] env[61857]: DEBUG nova.network.neutron [req-9ba979e4-32e4-4b79-9ab8-283d22259446 req-ccb34d8e-e5f2-4e18-84cb-b8e4ef839717 service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Refreshing network info cache for port d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 867.471311] env[61857]: DEBUG nova.compute.utils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.477030] env[61857]: DEBUG nova.compute.manager [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 867.477030] env[61857]: DEBUG nova.network.neutron [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 867.554234] env[61857]: DEBUG nova.policy [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a00d6256b5f4f78bc0fd160605123bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3eca39f9d1f244a6b73b33326f4eea7c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.571740] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52d41e1d-1ebf-3502-9650-033b01a80c04, 'name': SearchDatastore_Task, 'duration_secs': 0.013218} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.574438] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.574683] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.574931] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.575136] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.575332] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.576648] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e17a226-7248-4093-9b13-5bf85a08a845 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.584513] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.584765] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 867.585544] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26a4ff8c-66b5-4884-84a6-fb196a295ec1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.593137] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 867.593137] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52a4e613-0319-c052-86b8-e5933040654a" [ 867.593137] env[61857]: _type = "Task" [ 867.593137] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.601349] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a4e613-0319-c052-86b8-e5933040654a, 'name': SearchDatastore_Task} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.606443] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d5670dc-20e9-4fc4-9c9b-d92b2fa39368 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.612031] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 867.612031] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521a72d8-779a-86cd-b54d-1a1c89efddfb" [ 867.612031] env[61857]: _type = "Task" [ 867.612031] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.619333] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521a72d8-779a-86cd-b54d-1a1c89efddfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.771681] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquiring lock "refresh_cache-6957dc98-8c60-4fdd-83dd-be2f13825c6d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.771858] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquired lock "refresh_cache-6957dc98-8c60-4fdd-83dd-be2f13825c6d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.772034] env[61857]: DEBUG nova.network.neutron [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 867.856913] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951104, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.898697] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb29545-c287-419a-baa8-7decc7f67fea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.907250] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732cc5f0-7438-4185-8d85-8864f99cf1fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.939921] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0ea348-1423-4c3e-8f2a-ebcb8f6d826d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.949328] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8d1fa9-665f-4060-acd5-1f24733eb189 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.974797] env[61857]: DEBUG nova.compute.provider_tree [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.979761] env[61857]: DEBUG nova.compute.manager [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 867.995532] env[61857]: DEBUG nova.network.neutron [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Updated VIF entry in instance network info cache for port 7ede4f79-7428-48b0-a811-e8c064d224f0. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 867.995532] env[61857]: DEBUG nova.network.neutron [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Updating instance_info_cache with network_info: [{"id": "7ede4f79-7428-48b0-a811-e8c064d224f0", "address": "fa:16:3e:9e:cc:3a", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ede4f79-74", "ovs_interfaceid": "7ede4f79-7428-48b0-a811-e8c064d224f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.022792] env[61857]: DEBUG oslo_concurrency.lockutils [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "71920a62-090d-4df1-937e-87df3b043e28" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.023021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.023298] env[61857]: INFO nova.compute.manager [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Rebooting instance [ 868.128577] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521a72d8-779a-86cd-b54d-1a1c89efddfb, 'name': SearchDatastore_Task, 'duration_secs': 0.007926} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.129049] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.129608] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] ffa63de7-dd32-4908-92c5-755b3c0799ef/ffa63de7-dd32-4908-92c5-755b3c0799ef.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 868.130129] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3d459fe-efd9-434a-bf03-b004a1622e69 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.138955] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 868.138955] env[61857]: value = "task-951105" [ 868.138955] env[61857]: _type = "Task" [ 868.138955] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.148417] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951105, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.247786] env[61857]: DEBUG nova.network.neutron [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Successfully created port: d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 868.330433] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "c27f6f5b-6423-4eee-be53-dc48c93df893" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.330433] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "c27f6f5b-6423-4eee-be53-dc48c93df893" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.330433] env[61857]: INFO nova.compute.manager [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Rebooting instance [ 868.358825] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951104, 'name': ReconfigVM_Task, 'duration_secs': 0.814762} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.364031] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Reconfigured VM instance instance-0000004a to attach disk [datastore2] ddeed65b-9003-443f-8b2b-0756fbe7d234/ddeed65b-9003-443f-8b2b-0756fbe7d234.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.364031] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-60fde48f-5761-4f65-8aff-b47c6959b6fb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.366485] env[61857]: DEBUG nova.network.neutron [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 868.384010] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 868.384010] env[61857]: value = "task-951106" [ 868.384010] env[61857]: _type = "Task" [ 868.384010] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.402827] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951106, 'name': Rename_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.485462] env[61857]: DEBUG nova.scheduler.client.report [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.496355] env[61857]: DEBUG oslo_concurrency.lockutils [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] Releasing lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.496611] env[61857]: DEBUG nova.compute.manager [req-ed5289e1-96ce-45ab-9c55-31088d724097 req-8c63d916-f16e-459c-9e93-abdd7eb47ef4 service nova] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Received event network-vif-deleted-14eeb732-49e7-47b3-929e-85068b579c85 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 868.541639] env[61857]: DEBUG nova.network.neutron [req-9ba979e4-32e4-4b79-9ab8-283d22259446 req-ccb34d8e-e5f2-4e18-84cb-b8e4ef839717 service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Updated VIF entry in instance network info cache for port d1542c4d-1685-447b-a6c4-b56fad0ad098. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 868.542040] env[61857]: DEBUG nova.network.neutron [req-9ba979e4-32e4-4b79-9ab8-283d22259446 req-ccb34d8e-e5f2-4e18-84cb-b8e4ef839717 service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Updating instance_info_cache with network_info: [{"id": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "address": "fa:16:3e:20:7f:b7", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1542c4d-16", "ovs_interfaceid": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.558557] env[61857]: DEBUG oslo_concurrency.lockutils [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.558834] env[61857]: DEBUG oslo_concurrency.lockutils [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.559070] env[61857]: DEBUG nova.network.neutron [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 868.650046] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951105, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489971} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.650723] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] ffa63de7-dd32-4908-92c5-755b3c0799ef/ffa63de7-dd32-4908-92c5-755b3c0799ef.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 868.653594] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.653594] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-80c8dd15-14e6-4fc7-b286-d1fb3af55412 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.659211] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 868.659211] env[61857]: value = "task-951107" [ 868.659211] env[61857]: _type = "Task" [ 868.659211] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.674618] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951107, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.711238] env[61857]: DEBUG nova.network.neutron [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Updating instance_info_cache with network_info: [{"id": "feebcdf7-ef87-4705-90cf-ebb4415a92fc", "address": "fa:16:3e:5f:bf:c3", "network": {"id": "27c622a1-3c6c-4b63-b8e6-3aa5bbc9a44c", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-96124755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "192c67ad46484cf1b609342eb31962e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeebcdf7-ef", "ovs_interfaceid": "feebcdf7-ef87-4705-90cf-ebb4415a92fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.792638] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "3fb5d24b-1767-43d9-bed5-833121962adb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.792981] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "3fb5d24b-1767-43d9-bed5-833121962adb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.793222] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "3fb5d24b-1767-43d9-bed5-833121962adb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.793418] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "3fb5d24b-1767-43d9-bed5-833121962adb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.793593] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "3fb5d24b-1767-43d9-bed5-833121962adb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.796758] env[61857]: INFO nova.compute.manager [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Terminating instance [ 868.798061] env[61857]: DEBUG nova.compute.manager [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.798287] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 868.800056] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba6bd78-dda9-44bb-a162-bb9fc74c55d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.808657] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 868.809105] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-567e3f0b-3cc6-462c-bc34-66991b55bd85 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.816512] env[61857]: DEBUG oslo_vmware.api [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 868.816512] env[61857]: value = "task-951108" [ 868.816512] env[61857]: _type = "Task" [ 868.816512] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.824616] env[61857]: DEBUG oslo_vmware.api [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951108, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.849900] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.895557] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951106, 'name': Rename_Task, 'duration_secs': 0.267131} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.895557] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 868.895745] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-595fa18c-27e9-4428-8f1f-4853048bd53b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.902216] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 868.902216] env[61857]: value = "task-951109" [ 868.902216] env[61857]: _type = "Task" [ 868.902216] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.912253] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.993926] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.028s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.996957] env[61857]: DEBUG nova.compute.manager [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 868.999835] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.581s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.000846] env[61857]: INFO nova.compute.claims [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.022345] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.022655] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.022836] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.023402] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.023624] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.024355] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.025359] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.025607] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.025835] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.026037] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.026262] env[61857]: DEBUG nova.virt.hardware [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.027647] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f61b3d-9ca6-4888-89b4-4e661ac5fed3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.032872] env[61857]: INFO nova.scheduler.client.report [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Deleted allocations for instance 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5 [ 869.044478] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf51998-89cf-4dae-b044-a656d7cd0a32 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.048949] env[61857]: DEBUG oslo_concurrency.lockutils [req-9ba979e4-32e4-4b79-9ab8-283d22259446 req-ccb34d8e-e5f2-4e18-84cb-b8e4ef839717 service nova] Releasing lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.049410] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquired lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.049601] env[61857]: DEBUG nova.network.neutron [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 869.169014] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951107, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06451} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.169014] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.169871] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adeb958c-480f-4950-9417-bc3a246a6f7d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.194241] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] ffa63de7-dd32-4908-92c5-755b3c0799ef/ffa63de7-dd32-4908-92c5-755b3c0799ef.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.194563] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-174dd2b3-eb37-4eaa-b935-e624b0f4b1fa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.216118] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Releasing lock "refresh_cache-6957dc98-8c60-4fdd-83dd-be2f13825c6d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.216581] env[61857]: DEBUG nova.compute.manager [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Instance network_info: |[{"id": "feebcdf7-ef87-4705-90cf-ebb4415a92fc", "address": "fa:16:3e:5f:bf:c3", "network": {"id": "27c622a1-3c6c-4b63-b8e6-3aa5bbc9a44c", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-96124755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "192c67ad46484cf1b609342eb31962e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeebcdf7-ef", "ovs_interfaceid": "feebcdf7-ef87-4705-90cf-ebb4415a92fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 869.217920] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:bf:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f925dc8-2145-457e-a4d4-c07117356dd0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'feebcdf7-ef87-4705-90cf-ebb4415a92fc', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.226545] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Creating folder: Project (192c67ad46484cf1b609342eb31962e6). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 869.228900] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e50fa6e7-becf-4d22-869e-e9e079b0e584 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.230978] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 869.230978] env[61857]: value = "task-951110" [ 869.230978] env[61857]: _type = "Task" [ 869.230978] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.239942] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951110, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.241539] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Created folder: Project (192c67ad46484cf1b609342eb31962e6) in parent group-v214027. [ 869.241640] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Creating folder: Instances. Parent ref: group-v214141. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 869.241846] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-24c0f584-e161-433d-8a3d-6462a987a369 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.250942] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Created folder: Instances in parent group-v214141. [ 869.251222] env[61857]: DEBUG oslo.service.loopingcall [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.251433] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 869.251643] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4bafbb7c-46fc-44aa-b631-7542039de58b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.278354] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.278354] env[61857]: value = "task-951113" [ 869.278354] env[61857]: _type = "Task" [ 869.278354] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.286104] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951113, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.323919] env[61857]: DEBUG nova.compute.manager [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Received event network-vif-plugged-feebcdf7-ef87-4705-90cf-ebb4415a92fc {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 869.324227] env[61857]: DEBUG oslo_concurrency.lockutils [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] Acquiring lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.324466] env[61857]: DEBUG oslo_concurrency.lockutils [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] Lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.324639] env[61857]: DEBUG oslo_concurrency.lockutils [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] Lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.324877] env[61857]: DEBUG nova.compute.manager [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] No waiting events found dispatching network-vif-plugged-feebcdf7-ef87-4705-90cf-ebb4415a92fc {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.325162] env[61857]: WARNING nova.compute.manager [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Received unexpected event network-vif-plugged-feebcdf7-ef87-4705-90cf-ebb4415a92fc for instance with vm_state building and task_state spawning. [ 869.325397] env[61857]: DEBUG nova.compute.manager [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Received event network-changed-feebcdf7-ef87-4705-90cf-ebb4415a92fc {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 869.325518] env[61857]: DEBUG nova.compute.manager [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Refreshing instance network info cache due to event network-changed-feebcdf7-ef87-4705-90cf-ebb4415a92fc. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 869.325709] env[61857]: DEBUG oslo_concurrency.lockutils [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] Acquiring lock "refresh_cache-6957dc98-8c60-4fdd-83dd-be2f13825c6d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.325868] env[61857]: DEBUG oslo_concurrency.lockutils [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] Acquired lock "refresh_cache-6957dc98-8c60-4fdd-83dd-be2f13825c6d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.326012] env[61857]: DEBUG nova.network.neutron [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Refreshing network info cache for port feebcdf7-ef87-4705-90cf-ebb4415a92fc {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 869.331065] env[61857]: DEBUG oslo_vmware.api [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951108, 'name': PowerOffVM_Task, 'duration_secs': 0.359274} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.331394] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 869.331629] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 869.331827] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43bbb575-5466-4e16-8c54-1e6b575d1d6c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.416278] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951109, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.417275] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 869.417490] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 869.417677] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleting the datastore file [datastore2] 3fb5d24b-1767-43d9-bed5-833121962adb {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.417941] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec32a560-27e5-4e66-b0ca-a589abe8fa76 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.425232] env[61857]: DEBUG oslo_vmware.api [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 869.425232] env[61857]: value = "task-951115" [ 869.425232] env[61857]: _type = "Task" [ 869.425232] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.437029] env[61857]: DEBUG oslo_vmware.api [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951115, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.543910] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f51993e0-3833-4262-bfc4-0e49fe3592c6 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "36ca32ed-1ba2-40d0-93c4-086a26a3f9e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.772s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.635560] env[61857]: DEBUG nova.network.neutron [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance_info_cache with network_info: [{"id": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "address": "fa:16:3e:32:96:34", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c3f890-dd", "ovs_interfaceid": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.743769] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951110, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.789870] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951113, 'name': CreateVM_Task, 'duration_secs': 0.422459} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.790752] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 869.793964] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.793964] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.793964] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 869.793964] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9193099f-95a3-4ee5-bf84-4366433314d4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.798050] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for the task: (returnval){ [ 869.798050] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]529bbe87-b845-784f-ea6c-9a0a6a5b76f2" [ 869.798050] env[61857]: _type = "Task" [ 869.798050] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.810466] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529bbe87-b845-784f-ea6c-9a0a6a5b76f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.867900] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "interface-96ff157c-4471-4a12-ad99-0aafd6c2dce5-78436af1-f52d-459b-8460-184ce130e879" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.868239] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-96ff157c-4471-4a12-ad99-0aafd6c2dce5-78436af1-f52d-459b-8460-184ce130e879" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.889119] env[61857]: DEBUG nova.network.neutron [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Updating instance_info_cache with network_info: [{"id": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "address": "fa:16:3e:20:7f:b7", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1542c4d-16", "ovs_interfaceid": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.915733] env[61857]: DEBUG oslo_vmware.api [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951109, 'name': PowerOnVM_Task, 'duration_secs': 0.520548} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.916015] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 869.916442] env[61857]: INFO nova.compute.manager [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Took 8.83 seconds to spawn the instance on the hypervisor. [ 869.916641] env[61857]: DEBUG nova.compute.manager [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.917473] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d55682-b7d1-4dc1-b969-ea8aed79b088 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.939841] env[61857]: DEBUG oslo_vmware.api [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951115, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216884} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.939841] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.939841] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 869.940141] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 869.940359] env[61857]: INFO nova.compute.manager [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Took 1.14 seconds to destroy the instance on the hypervisor. [ 869.941088] env[61857]: DEBUG oslo.service.loopingcall [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.941088] env[61857]: DEBUG nova.compute.manager [-] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.941088] env[61857]: DEBUG nova.network.neutron [-] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 870.138033] env[61857]: DEBUG oslo_concurrency.lockutils [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.144285] env[61857]: DEBUG nova.compute.manager [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.149613] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09658dff-11c3-43c7-ade3-1537af1fff19 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.254451] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951110, 'name': ReconfigVM_Task, 'duration_secs': 0.568363} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.257434] env[61857]: DEBUG nova.network.neutron [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Updated VIF entry in instance network info cache for port feebcdf7-ef87-4705-90cf-ebb4415a92fc. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 870.257434] env[61857]: DEBUG nova.network.neutron [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Updating instance_info_cache with network_info: [{"id": "feebcdf7-ef87-4705-90cf-ebb4415a92fc", "address": "fa:16:3e:5f:bf:c3", "network": {"id": "27c622a1-3c6c-4b63-b8e6-3aa5bbc9a44c", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-96124755-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "192c67ad46484cf1b609342eb31962e6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeebcdf7-ef", "ovs_interfaceid": "feebcdf7-ef87-4705-90cf-ebb4415a92fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.257434] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Reconfigured VM instance instance-0000004b to attach disk [datastore2] ffa63de7-dd32-4908-92c5-755b3c0799ef/ffa63de7-dd32-4908-92c5-755b3c0799ef.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 870.260827] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a1f8c45-1253-4bb0-aff9-3f76a25a95ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.272689] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 870.272689] env[61857]: value = "task-951116" [ 870.272689] env[61857]: _type = "Task" [ 870.272689] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.284900] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951116, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.313025] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529bbe87-b845-784f-ea6c-9a0a6a5b76f2, 'name': SearchDatastore_Task, 'duration_secs': 0.011036} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.313394] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.313666] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.313902] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.314089] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.314287] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.314556] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04eaac16-8a84-41fb-a9c4-0b6b6dff1ff6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.319539] env[61857]: DEBUG nova.compute.manager [req-2ea31792-9ca8-485f-b90f-519e17187a46 req-689544a1-d20e-40e2-b1a8-652a81f3574d service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Received event network-vif-plugged-d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 870.319789] env[61857]: DEBUG oslo_concurrency.lockutils [req-2ea31792-9ca8-485f-b90f-519e17187a46 req-689544a1-d20e-40e2-b1a8-652a81f3574d service nova] Acquiring lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.320049] env[61857]: DEBUG oslo_concurrency.lockutils [req-2ea31792-9ca8-485f-b90f-519e17187a46 req-689544a1-d20e-40e2-b1a8-652a81f3574d service nova] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.320324] env[61857]: DEBUG oslo_concurrency.lockutils [req-2ea31792-9ca8-485f-b90f-519e17187a46 req-689544a1-d20e-40e2-b1a8-652a81f3574d service nova] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.320558] env[61857]: DEBUG nova.compute.manager [req-2ea31792-9ca8-485f-b90f-519e17187a46 req-689544a1-d20e-40e2-b1a8-652a81f3574d service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] No waiting events found dispatching network-vif-plugged-d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.320758] env[61857]: WARNING nova.compute.manager [req-2ea31792-9ca8-485f-b90f-519e17187a46 req-689544a1-d20e-40e2-b1a8-652a81f3574d service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Received unexpected event network-vif-plugged-d9f80f94-9be0-4d62-863f-7562c6c9e63e for instance with vm_state building and task_state spawning. [ 870.325478] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.325478] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 870.326674] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93f1fc11-f186-450a-b588-96bd3ec3a029 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.334444] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for the task: (returnval){ [ 870.334444] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ee8aac-ea15-4b9d-7927-dabf97c6c005" [ 870.334444] env[61857]: _type = "Task" [ 870.334444] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.351340] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ee8aac-ea15-4b9d-7927-dabf97c6c005, 'name': SearchDatastore_Task, 'duration_secs': 0.011367} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.351853] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cac447a-92cd-4411-b55b-c0b537ee1cff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.359607] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for the task: (returnval){ [ 870.359607] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]528048ee-c856-355a-5d5b-d2fdaaa3b6d0" [ 870.359607] env[61857]: _type = "Task" [ 870.359607] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.368500] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528048ee-c856-355a-5d5b-d2fdaaa3b6d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.373714] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.373898] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.374728] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637c8706-4a9c-43af-b856-935946671fae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.398723] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Releasing lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.401425] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5ed4ea-4e7a-4e2d-9b29-fc9f63d1d9ef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.404284] env[61857]: DEBUG nova.compute.manager [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.405074] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68bcbf6-320a-4187-acdc-89ee410c745d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.437797] env[61857]: DEBUG nova.compute.manager [req-2944fc17-6cdd-43e7-8d2f-707b68e2aa84 req-01d35bfc-57b8-4224-b50b-547946cd564e service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Received event network-vif-deleted-472be223-da36-4834-9f18-05437505e7b8 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 870.437962] env[61857]: INFO nova.compute.manager [req-2944fc17-6cdd-43e7-8d2f-707b68e2aa84 req-01d35bfc-57b8-4224-b50b-547946cd564e service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Neutron deleted interface 472be223-da36-4834-9f18-05437505e7b8; detaching it from the instance and deleting it from the info cache [ 870.438159] env[61857]: DEBUG nova.network.neutron [req-2944fc17-6cdd-43e7-8d2f-707b68e2aa84 req-01d35bfc-57b8-4224-b50b-547946cd564e service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.448363] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Reconfiguring VM to detach interface {{(pid=61857) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 870.453401] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-664b5173-7556-49b9-935d-99b7379dc7d8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.466929] env[61857]: INFO nova.compute.manager [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Took 36.04 seconds to build instance. [ 870.476480] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 870.476480] env[61857]: value = "task-951117" [ 870.476480] env[61857]: _type = "Task" [ 870.476480] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.487275] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.593024] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4997cd-038f-498e-ba53-97abc11459c3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.601075] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988c074d-d2bb-4c78-b6b7-9da5f5a6e8d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.639173] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6320d18-1211-437d-aded-982eb68a4a8a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.645394] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccc7459-d08f-4f8e-9197-6b2f5ec4704f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.666785] env[61857]: DEBUG nova.compute.provider_tree [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.764182] env[61857]: DEBUG oslo_concurrency.lockutils [req-67166e63-20b7-403f-bf56-d2bc46b303a4 req-42f999a6-ca73-4150-b97a-197445eb50f7 service nova] Releasing lock "refresh_cache-6957dc98-8c60-4fdd-83dd-be2f13825c6d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.538098] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "ddeed65b-9003-443f-8b2b-0756fbe7d234" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.541881] env[61857]: DEBUG nova.network.neutron [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Successfully updated port: d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.542902] env[61857]: DEBUG nova.network.neutron [-] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.548054] env[61857]: DEBUG oslo_concurrency.lockutils [None req-870e0a53-c74b-4707-98b1-0dd44644a629 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "ddeed65b-9003-443f-8b2b-0756fbe7d234" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.038s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.548054] env[61857]: DEBUG nova.scheduler.client.report [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.553018] env[61857]: DEBUG nova.compute.manager [req-fb056a1f-6878-49cd-ac73-6341fc09865f req-9c3e84fd-f293-4948-8e4d-868c4ab9da4c service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Received event network-changed-d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 871.553018] env[61857]: DEBUG nova.compute.manager [req-fb056a1f-6878-49cd-ac73-6341fc09865f req-9c3e84fd-f293-4948-8e4d-868c4ab9da4c service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Refreshing instance network info cache due to event network-changed-d9f80f94-9be0-4d62-863f-7562c6c9e63e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 871.553018] env[61857]: DEBUG oslo_concurrency.lockutils [req-fb056a1f-6878-49cd-ac73-6341fc09865f req-9c3e84fd-f293-4948-8e4d-868c4ab9da4c service nova] Acquiring lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.553018] env[61857]: DEBUG oslo_concurrency.lockutils [req-fb056a1f-6878-49cd-ac73-6341fc09865f req-9c3e84fd-f293-4948-8e4d-868c4ab9da4c service nova] Acquired lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.553194] env[61857]: DEBUG nova.network.neutron [req-fb056a1f-6878-49cd-ac73-6341fc09865f req-9c3e84fd-f293-4948-8e4d-868c4ab9da4c service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Refreshing network info cache for port d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.555582] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951116, 'name': Rename_Task, 'duration_secs': 0.26788} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.558886] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a7ce02cb-4202-409d-9468-36690b8bf8a3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.560861] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "ddeed65b-9003-443f-8b2b-0756fbe7d234" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.023s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.561140] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "ddeed65b-9003-443f-8b2b-0756fbe7d234-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.561392] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "ddeed65b-9003-443f-8b2b-0756fbe7d234-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.561580] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "ddeed65b-9003-443f-8b2b-0756fbe7d234-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.566111] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e36f4e-1a62-4f8e-baa1-40b31dd9623e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.568823] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 871.586108] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3c7952a-a34f-4a77-a77f-c631f65e91b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.586108] env[61857]: INFO nova.compute.manager [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Terminating instance [ 871.586108] env[61857]: DEBUG nova.compute.manager [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.586108] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 871.586108] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420f1722-aaae-4ecd-b1fd-1cce824ba4e4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.595307] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.595743] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528048ee-c856-355a-5d5b-d2fdaaa3b6d0, 'name': SearchDatastore_Task, 'duration_secs': 0.019976} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.596150] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Doing hard reboot of VM {{(pid=61857) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 871.598383] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.598598] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 6957dc98-8c60-4fdd-83dd-be2f13825c6d/6957dc98-8c60-4fdd-83dd-be2f13825c6d.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 871.601663] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50b9239-76e8-4853-91fd-b068cbb3088c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.612371] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-0e3e9078-3116-4c40-bee0-77549a3c7d58 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.616936] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79c3800d-b9a5-48c7-9757-61984679b050 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.620103] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 871.620103] env[61857]: value = "task-951118" [ 871.620103] env[61857]: _type = "Task" [ 871.620103] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.630852] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 871.633428] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f78af1d-c9a6-4eb2-b2b9-cac6ab48dc14 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.633661] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for the task: (returnval){ [ 871.633661] env[61857]: value = "task-951120" [ 871.633661] env[61857]: _type = "Task" [ 871.633661] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.634011] env[61857]: DEBUG oslo_vmware.api [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 871.634011] env[61857]: value = "task-951119" [ 871.634011] env[61857]: _type = "Task" [ 871.634011] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.639348] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951118, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.663531] env[61857]: DEBUG nova.compute.manager [req-2944fc17-6cdd-43e7-8d2f-707b68e2aa84 req-01d35bfc-57b8-4224-b50b-547946cd564e service nova] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Detach interface failed, port_id=472be223-da36-4834-9f18-05437505e7b8, reason: Instance 3fb5d24b-1767-43d9-bed5-833121962adb could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 871.664709] env[61857]: DEBUG oslo_vmware.api [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 871.664709] env[61857]: value = "task-951121" [ 871.664709] env[61857]: _type = "Task" [ 871.664709] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.675596] env[61857]: DEBUG oslo_vmware.api [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951119, 'name': ResetVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.675596] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951120, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.682581] env[61857]: DEBUG oslo_vmware.api [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.059457] env[61857]: INFO nova.compute.manager [-] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Took 2.12 seconds to deallocate network for instance. [ 872.059948] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.060716] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.061218] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6cfd3fd-96bc-4a99-92f1-6f1a59fa524a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.071256] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Doing hard reboot of VM {{(pid=61857) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 872.071760] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-bebc7a89-a436-44f5-9f61-6cae56240408 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.080807] env[61857]: DEBUG oslo_vmware.api [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 872.080807] env[61857]: value = "task-951122" [ 872.080807] env[61857]: _type = "Task" [ 872.080807] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.090143] env[61857]: DEBUG oslo_vmware.api [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951122, 'name': ResetVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.129336] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.130s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.129884] env[61857]: DEBUG nova.compute.manager [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.137833] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 22.670s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.139385] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951118, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.153157] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951120, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453293} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.156899] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 6957dc98-8c60-4fdd-83dd-be2f13825c6d/6957dc98-8c60-4fdd-83dd-be2f13825c6d.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 872.157159] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.157505] env[61857]: DEBUG oslo_vmware.api [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951119, 'name': ResetVM_Task, 'duration_secs': 0.109805} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.157753] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c11ea653-830d-4462-be2e-46f751c55cf6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.160528] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Did hard reboot of VM {{(pid=61857) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 872.160769] env[61857]: DEBUG nova.compute.manager [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.161589] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefde11e-9bff-4a4d-965a-f206b5608fef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.176700] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for the task: (returnval){ [ 872.176700] env[61857]: value = "task-951123" [ 872.176700] env[61857]: _type = "Task" [ 872.176700] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.183476] env[61857]: DEBUG oslo_vmware.api [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951121, 'name': PowerOffVM_Task, 'duration_secs': 0.236089} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.184435] env[61857]: DEBUG nova.network.neutron [req-fb056a1f-6878-49cd-ac73-6341fc09865f req-9c3e84fd-f293-4948-8e4d-868c4ab9da4c service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 872.186767] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 872.186879] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 872.187196] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-761c2d85-6559-4e6e-b578-61d6c6e4ae19 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.192323] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951123, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.282137] env[61857]: DEBUG nova.network.neutron [req-fb056a1f-6878-49cd-ac73-6341fc09865f req-9c3e84fd-f293-4948-8e4d-868c4ab9da4c service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.286977] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 872.287343] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 872.287513] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleting the datastore file [datastore2] ddeed65b-9003-443f-8b2b-0756fbe7d234 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.288087] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-319a1d68-624f-4b8e-8c68-337faae3ffb5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.297757] env[61857]: DEBUG oslo_vmware.api [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 872.297757] env[61857]: value = "task-951125" [ 872.297757] env[61857]: _type = "Task" [ 872.297757] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.307734] env[61857]: DEBUG oslo_vmware.api [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951125, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.559686] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.573072] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.590923] env[61857]: DEBUG oslo_vmware.api [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951122, 'name': ResetVM_Task, 'duration_secs': 0.095037} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.591253] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Did hard reboot of VM {{(pid=61857) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 872.591435] env[61857]: DEBUG nova.compute.manager [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.592224] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b95979-fb70-41fa-92d1-c943b8efe56f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.632487] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951118, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.639192] env[61857]: DEBUG nova.compute.utils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.640587] env[61857]: DEBUG nova.compute.manager [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.640757] env[61857]: DEBUG nova.network.neutron [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 872.686800] env[61857]: DEBUG oslo_concurrency.lockutils [None req-098222b7-04a2-409d-a0f8-ed731d1933f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.664s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.696276] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951123, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073277} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.697780] env[61857]: DEBUG nova.policy [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dddec4385c44cef8508a887e1aaee9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3fb60d26b96c453b8eb4aee57b040d7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.699320] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 872.700802] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42d4dd8-b7a6-4923-a2f5-d7f81996d868 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.729567] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 6957dc98-8c60-4fdd-83dd-be2f13825c6d/6957dc98-8c60-4fdd-83dd-be2f13825c6d.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.732608] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2827f97-e49a-4338-ae60-0e3889066e9f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.756314] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for the task: (returnval){ [ 872.756314] env[61857]: value = "task-951126" [ 872.756314] env[61857]: _type = "Task" [ 872.756314] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.769093] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951126, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.781274] env[61857]: DEBUG oslo_concurrency.lockutils [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.781584] env[61857]: DEBUG oslo_concurrency.lockutils [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.781801] env[61857]: DEBUG oslo_concurrency.lockutils [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.781985] env[61857]: DEBUG oslo_concurrency.lockutils [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.782174] env[61857]: DEBUG oslo_concurrency.lockutils [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.784894] env[61857]: INFO nova.compute.manager [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Terminating instance [ 872.786626] env[61857]: DEBUG oslo_concurrency.lockutils [req-fb056a1f-6878-49cd-ac73-6341fc09865f req-9c3e84fd-f293-4948-8e4d-868c4ab9da4c service nova] Releasing lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.790470] env[61857]: DEBUG nova.compute.manager [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 872.790692] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 872.790978] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.791207] env[61857]: DEBUG nova.network.neutron [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 872.792789] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558d4bf6-14e3-450b-a2a6-db495be7d82d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.806321] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 872.806931] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94be6b23-dacb-4ebf-98d2-1246b4630195 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.813497] env[61857]: DEBUG oslo_vmware.api [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951125, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155744} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.815122] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.815990] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 872.816338] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 872.816537] env[61857]: INFO nova.compute.manager [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Took 1.24 seconds to destroy the instance on the hypervisor. [ 872.816793] env[61857]: DEBUG oslo.service.loopingcall [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.817103] env[61857]: DEBUG oslo_vmware.api [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 872.817103] env[61857]: value = "task-951127" [ 872.817103] env[61857]: _type = "Task" [ 872.817103] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.819952] env[61857]: DEBUG nova.compute.manager [-] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 872.820063] env[61857]: DEBUG nova.network.neutron [-] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 872.832205] env[61857]: DEBUG oslo_vmware.api [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951127, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.061295] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.106673] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b864315f-d432-4b02-8ac5-6f2beab1d561 tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "c27f6f5b-6423-4eee-be53-dc48c93df893" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.778s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.123815] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.124079] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.141787] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951118, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.146276] env[61857]: DEBUG nova.compute.manager [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.156923] env[61857]: DEBUG nova.network.neutron [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Successfully created port: 6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.185099] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ac895e-dcdf-41a3-bf89-d2e70a72eb7b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.194013] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8e9d33-e846-4d48-b0b0-ca45bb861453 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.231025] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30e50a4-c23e-4ad4-a606-bfc0a3e5d9a4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.244070] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d35490a-8357-4bfe-af23-69936f1db7fb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.256142] env[61857]: DEBUG nova.compute.provider_tree [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.266522] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951126, 'name': ReconfigVM_Task, 'duration_secs': 0.310534} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.267286] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 6957dc98-8c60-4fdd-83dd-be2f13825c6d/6957dc98-8c60-4fdd-83dd-be2f13825c6d.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.267928] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-516b5dac-bace-417c-a42e-4a0385422cb6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.274672] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for the task: (returnval){ [ 873.274672] env[61857]: value = "task-951128" [ 873.274672] env[61857]: _type = "Task" [ 873.274672] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.284791] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951128, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.312365] env[61857]: DEBUG nova.compute.manager [req-7b2ebaa6-af45-4ffa-948b-397acf709e9f req-1f02eca5-c241-4419-9232-15300bd637eb service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Received event network-vif-deleted-ff5c225b-875a-4492-b57d-a91e1063a36e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 873.312599] env[61857]: INFO nova.compute.manager [req-7b2ebaa6-af45-4ffa-948b-397acf709e9f req-1f02eca5-c241-4419-9232-15300bd637eb service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Neutron deleted interface ff5c225b-875a-4492-b57d-a91e1063a36e; detaching it from the instance and deleting it from the info cache [ 873.312818] env[61857]: DEBUG nova.network.neutron [req-7b2ebaa6-af45-4ffa-948b-397acf709e9f req-1f02eca5-c241-4419-9232-15300bd637eb service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.332932] env[61857]: DEBUG oslo_vmware.api [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951127, 'name': PowerOffVM_Task, 'duration_secs': 0.261849} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.334278] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 873.334278] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 873.334278] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4aaac99c-9afe-48ee-a103-3a07a6e680c2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.342835] env[61857]: DEBUG nova.network.neutron [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 873.406334] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 873.406582] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 873.406810] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Deleting the datastore file [datastore1] 6c4c1276-ae80-4f37-9ef9-4872dd415d24 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.407180] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d44b2bd3-37e6-4543-9f14-ff7ba43ecee3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.414818] env[61857]: DEBUG oslo_vmware.api [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 873.414818] env[61857]: value = "task-951130" [ 873.414818] env[61857]: _type = "Task" [ 873.414818] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.423299] env[61857]: DEBUG oslo_vmware.api [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951130, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.498616] env[61857]: DEBUG nova.network.neutron [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updating instance_info_cache with network_info: [{"id": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "address": "fa:16:3e:94:19:6b", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9f80f94-9b", "ovs_interfaceid": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.560675] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.630767] env[61857]: DEBUG nova.compute.manager [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.638383] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951118, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.735365] env[61857]: DEBUG nova.network.neutron [-] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.762481] env[61857]: DEBUG nova.scheduler.client.report [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.788800] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951128, 'name': Rename_Task, 'duration_secs': 0.152445} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.789125] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 873.789397] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-143fa18a-75f7-4ad1-9ddc-edcd0938b83e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.798103] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for the task: (returnval){ [ 873.798103] env[61857]: value = "task-951131" [ 873.798103] env[61857]: _type = "Task" [ 873.798103] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.811895] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951131, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.818521] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17a54d25-391e-4d79-956e-83529d9ea77b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.828501] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c924760-e7ab-4d34-a802-f2cf99efe28e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.875841] env[61857]: DEBUG nova.compute.manager [req-7b2ebaa6-af45-4ffa-948b-397acf709e9f req-1f02eca5-c241-4419-9232-15300bd637eb service nova] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Detach interface failed, port_id=ff5c225b-875a-4492-b57d-a91e1063a36e, reason: Instance ddeed65b-9003-443f-8b2b-0756fbe7d234 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 873.926676] env[61857]: DEBUG oslo_vmware.api [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951130, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153582} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.926874] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.926994] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 873.927613] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 873.927613] env[61857]: INFO nova.compute.manager [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Took 1.14 seconds to destroy the instance on the hypervisor. [ 873.927613] env[61857]: DEBUG oslo.service.loopingcall [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.927787] env[61857]: DEBUG nova.compute.manager [-] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.927883] env[61857]: DEBUG nova.network.neutron [-] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 874.001069] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.001388] env[61857]: DEBUG nova.compute.manager [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Instance network_info: |[{"id": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "address": "fa:16:3e:94:19:6b", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9f80f94-9b", "ovs_interfaceid": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 874.001939] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:19:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6edb8eae-1113-49d0-84f7-9fd9f82b26fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9f80f94-9be0-4d62-863f-7562c6c9e63e', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 874.011319] env[61857]: DEBUG oslo.service.loopingcall [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.011664] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 874.012033] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04b24664-d32e-408f-8652-e806774ad053 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.035421] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 874.035421] env[61857]: value = "task-951132" [ 874.035421] env[61857]: _type = "Task" [ 874.035421] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.048745] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951132, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.058267] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.134932] env[61857]: DEBUG oslo_vmware.api [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951118, 'name': PowerOnVM_Task, 'duration_secs': 2.230005} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.135451] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 874.135451] env[61857]: INFO nova.compute.manager [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Took 10.47 seconds to spawn the instance on the hypervisor. [ 874.135654] env[61857]: DEBUG nova.compute.manager [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.136375] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81e066a-bc93-497d-97ff-22cf834c5665 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.156851] env[61857]: DEBUG nova.compute.manager [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 874.160168] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.186153] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.186529] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.186712] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.186906] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.187075] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.187233] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.187452] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.187617] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.187787] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.187954] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.188185] env[61857]: DEBUG nova.virt.hardware [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.189405] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74bd5831-a658-4518-8dca-44c3d1bff5a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.199948] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839cccb6-ce73-4274-9971-ffb91a1be48e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.237851] env[61857]: INFO nova.compute.manager [-] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Took 1.42 seconds to deallocate network for instance. [ 874.309503] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951131, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.546825] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951132, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.558807] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.658923] env[61857]: INFO nova.compute.manager [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Took 39.73 seconds to build instance. [ 874.745378] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.775851] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.638s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.778696] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.354s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.778877] env[61857]: DEBUG nova.objects.instance [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61857) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 874.794561] env[61857]: DEBUG nova.compute.manager [req-11ac9ca2-db0b-426b-9d68-c32c1797188a req-e21fc000-480f-442d-a6c4-739f479951a1 service nova] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Received event network-vif-plugged-6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 874.794782] env[61857]: DEBUG oslo_concurrency.lockutils [req-11ac9ca2-db0b-426b-9d68-c32c1797188a req-e21fc000-480f-442d-a6c4-739f479951a1 service nova] Acquiring lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.794989] env[61857]: DEBUG oslo_concurrency.lockutils [req-11ac9ca2-db0b-426b-9d68-c32c1797188a req-e21fc000-480f-442d-a6c4-739f479951a1 service nova] Lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.795176] env[61857]: DEBUG oslo_concurrency.lockutils [req-11ac9ca2-db0b-426b-9d68-c32c1797188a req-e21fc000-480f-442d-a6c4-739f479951a1 service nova] Lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.795362] env[61857]: DEBUG nova.compute.manager [req-11ac9ca2-db0b-426b-9d68-c32c1797188a req-e21fc000-480f-442d-a6c4-739f479951a1 service nova] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] No waiting events found dispatching network-vif-plugged-6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.795528] env[61857]: WARNING nova.compute.manager [req-11ac9ca2-db0b-426b-9d68-c32c1797188a req-e21fc000-480f-442d-a6c4-739f479951a1 service nova] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Received unexpected event network-vif-plugged-6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2 for instance with vm_state building and task_state spawning. [ 874.811141] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951131, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.822506] env[61857]: DEBUG nova.network.neutron [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Successfully updated port: 6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.889204] env[61857]: DEBUG nova.network.neutron [-] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.048958] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951132, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.059552] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.161404] env[61857]: DEBUG oslo_concurrency.lockutils [None req-71cf24cc-5d60-45e1-b3cb-ee89db813e2f tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.239s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.271697] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "c27f6f5b-6423-4eee-be53-dc48c93df893" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.271961] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "c27f6f5b-6423-4eee-be53-dc48c93df893" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.272200] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "c27f6f5b-6423-4eee-be53-dc48c93df893-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.272392] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "c27f6f5b-6423-4eee-be53-dc48c93df893-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.272568] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "c27f6f5b-6423-4eee-be53-dc48c93df893-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.274938] env[61857]: INFO nova.compute.manager [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Terminating instance [ 875.277217] env[61857]: DEBUG nova.compute.manager [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 875.277426] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 875.279136] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030f7e0c-1897-4e42-80ba-fd772749eb6d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.292707] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 875.293629] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-baaa7dc7-206b-4096-90e0-88ee19961b8b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.305433] env[61857]: DEBUG oslo_vmware.api [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 875.305433] env[61857]: value = "task-951133" [ 875.305433] env[61857]: _type = "Task" [ 875.305433] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.316454] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951131, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.322996] env[61857]: DEBUG oslo_vmware.api [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951133, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.325851] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquiring lock "refresh_cache-0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.326032] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquired lock "refresh_cache-0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.326307] env[61857]: DEBUG nova.network.neutron [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.380338] env[61857]: INFO nova.scheduler.client.report [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleted allocation for migration f0272d66-8a8c-48a8-aad1-69a9053fe892 [ 875.392051] env[61857]: INFO nova.compute.manager [-] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Took 1.46 seconds to deallocate network for instance. [ 875.548427] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951132, 'name': CreateVM_Task, 'duration_secs': 1.19118} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.548427] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 875.549195] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.549335] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.549821] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.549918] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a927ea6-6a77-44b9-b91e-80902d62f9e9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.558326] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 875.558326] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52f9ca50-cf48-0cac-cc4a-fa1750b2f642" [ 875.558326] env[61857]: _type = "Task" [ 875.558326] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.561555] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.569174] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f9ca50-cf48-0cac-cc4a-fa1750b2f642, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.581191] env[61857]: DEBUG nova.compute.manager [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Received event network-vif-deleted-a930076b-3950-47c6-a108-dc309a22bda4 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 875.581469] env[61857]: DEBUG nova.compute.manager [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Received event network-changed-d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 875.581585] env[61857]: DEBUG nova.compute.manager [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Refreshing instance network info cache due to event network-changed-d1542c4d-1685-447b-a6c4-b56fad0ad098. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 875.581775] env[61857]: DEBUG oslo_concurrency.lockutils [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] Acquiring lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.581924] env[61857]: DEBUG oslo_concurrency.lockutils [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] Acquired lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.582228] env[61857]: DEBUG nova.network.neutron [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Refreshing network info cache for port d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 875.793513] env[61857]: DEBUG oslo_concurrency.lockutils [None req-dfb5f1d1-4c21-4c70-a9e9-4b63d09d7a7b tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.794653] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.337s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.794899] env[61857]: DEBUG nova.objects.instance [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lazy-loading 'resources' on Instance uuid ad89cb5f-44d0-45c9-92a5-7f7ff3138e21 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.816364] env[61857]: DEBUG oslo_vmware.api [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951131, 'name': PowerOnVM_Task, 'duration_secs': 1.550916} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.817099] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 875.817347] env[61857]: INFO nova.compute.manager [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Took 9.55 seconds to spawn the instance on the hypervisor. [ 875.817532] env[61857]: DEBUG nova.compute.manager [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 875.818355] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ace4d4e-0383-4b37-92e0-18e86b56dc0d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.824772] env[61857]: DEBUG oslo_vmware.api [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951133, 'name': PowerOffVM_Task, 'duration_secs': 0.232655} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.825447] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 875.825651] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 875.825903] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d263a362-dec6-49c6-a815-785852905eac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.874220] env[61857]: DEBUG nova.network.neutron [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 875.893639] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73f1b240-411a-48e7-9c2b-335d60eb42f7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 29.328s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.900988] env[61857]: DEBUG oslo_concurrency.lockutils [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.931209] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 875.931507] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 875.931702] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Deleting the datastore file [datastore2] c27f6f5b-6423-4eee-be53-dc48c93df893 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.932412] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8ebf503-766c-48b8-a845-a68715edb5f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.939887] env[61857]: DEBUG oslo_vmware.api [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 875.939887] env[61857]: value = "task-951135" [ 875.939887] env[61857]: _type = "Task" [ 875.939887] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.953236] env[61857]: DEBUG oslo_vmware.api [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951135, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.064660] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.074153] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f9ca50-cf48-0cac-cc4a-fa1750b2f642, 'name': SearchDatastore_Task, 'duration_secs': 0.012979} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.074504] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.074810] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.075047] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.075200] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.075392] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.075699] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-912300d6-35da-420a-8fc1-4c978db923c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.087129] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.087129] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 876.089417] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7022801a-23ee-4392-ac90-00857193fad5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.100016] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 876.100016] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5237f03e-65f7-251d-9fe5-56c33a1bdfec" [ 876.100016] env[61857]: _type = "Task" [ 876.100016] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.107337] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5237f03e-65f7-251d-9fe5-56c33a1bdfec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.165654] env[61857]: DEBUG nova.network.neutron [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Updating instance_info_cache with network_info: [{"id": "6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2", "address": "fa:16:3e:3e:cf:79", "network": {"id": "2b3f7088-e451-4d24-bda6-faaf0cac6afa", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-851060206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fb60d26b96c453b8eb4aee57b040d7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a874c214-8cdf-4a41-a718-84262b2a28d8", "external-id": "cl2-zone-726", "segmentation_id": 726, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6dc5c15d-c7", "ovs_interfaceid": "6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.344976] env[61857]: INFO nova.compute.manager [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Took 37.41 seconds to build instance. [ 876.418744] env[61857]: DEBUG nova.network.neutron [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Updated VIF entry in instance network info cache for port d1542c4d-1685-447b-a6c4-b56fad0ad098. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 876.419217] env[61857]: DEBUG nova.network.neutron [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Updating instance_info_cache with network_info: [{"id": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "address": "fa:16:3e:20:7f:b7", "network": {"id": "e5f928e2-6ad9-49ca-92a4-66e31a3b8f4f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-345545559-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e3643abae2e40a99059603131e0b7b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1542c4d-16", "ovs_interfaceid": "d1542c4d-1685-447b-a6c4-b56fad0ad098", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.455630] env[61857]: DEBUG oslo_vmware.api [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951135, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174674} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.455770] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 876.455917] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 876.456312] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 876.456570] env[61857]: INFO nova.compute.manager [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Took 1.18 seconds to destroy the instance on the hypervisor. [ 876.456823] env[61857]: DEBUG oslo.service.loopingcall [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.459636] env[61857]: DEBUG nova.compute.manager [-] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 876.459745] env[61857]: DEBUG nova.network.neutron [-] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 876.563046] env[61857]: DEBUG oslo_vmware.api [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951117, 'name': ReconfigVM_Task, 'duration_secs': 5.778769} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.565278] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.565503] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Reconfigured VM to detach interface {{(pid=61857) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 876.610115] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5237f03e-65f7-251d-9fe5-56c33a1bdfec, 'name': SearchDatastore_Task, 'duration_secs': 0.010892} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.618378] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e19433f-f8c6-4153-a977-9edb5c578aa9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.625940] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 876.625940] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e0189f-c496-b91f-e955-ce4ff7059712" [ 876.625940] env[61857]: _type = "Task" [ 876.625940] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.639135] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e0189f-c496-b91f-e955-ce4ff7059712, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.668494] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Releasing lock "refresh_cache-0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.668851] env[61857]: DEBUG nova.compute.manager [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Instance network_info: |[{"id": "6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2", "address": "fa:16:3e:3e:cf:79", "network": {"id": "2b3f7088-e451-4d24-bda6-faaf0cac6afa", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-851060206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fb60d26b96c453b8eb4aee57b040d7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a874c214-8cdf-4a41-a718-84262b2a28d8", "external-id": "cl2-zone-726", "segmentation_id": 726, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6dc5c15d-c7", "ovs_interfaceid": "6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 876.669325] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:cf:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a874c214-8cdf-4a41-a718-84262b2a28d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.680128] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Creating folder: Project (3fb60d26b96c453b8eb4aee57b040d7e). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 876.680245] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-68181022-68c5-4bb4-8a5b-8da2052368a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.696227] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Created folder: Project (3fb60d26b96c453b8eb4aee57b040d7e) in parent group-v214027. [ 876.696510] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Creating folder: Instances. Parent ref: group-v214145. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 876.696826] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7c68d30f-bf69-4311-a3fd-8fac29ba2f54 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.711329] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Created folder: Instances in parent group-v214145. [ 876.712397] env[61857]: DEBUG oslo.service.loopingcall [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.712397] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 876.712397] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f9652c57-f21f-480b-b0ea-4af8b5a288ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.743524] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.743524] env[61857]: value = "task-951138" [ 876.743524] env[61857]: _type = "Task" [ 876.743524] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.755366] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951138, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.770788] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44457a9-836e-4372-89ca-0f852d897e93 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.785958] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83bc7c67-2a35-4443-9672-fa558df3897c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.822688] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4ce12e-6d1e-4220-9615-21c698f81be2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.831820] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c304369-4087-4e4a-a366-573c858cb8c5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.850715] env[61857]: DEBUG oslo_concurrency.lockutils [None req-68b2498c-84a1-42e9-a7ed-d65bd60a269a tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.282s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.851559] env[61857]: DEBUG nova.compute.provider_tree [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.922750] env[61857]: DEBUG oslo_concurrency.lockutils [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] Releasing lock "refresh_cache-c27f6f5b-6423-4eee-be53-dc48c93df893" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.923877] env[61857]: DEBUG nova.compute.manager [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Received event network-changed-25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.924035] env[61857]: DEBUG nova.compute.manager [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Refreshing instance network info cache due to event network-changed-25a85668-66ff-4cf6-8cef-68b4d1a2e828. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 876.924850] env[61857]: DEBUG oslo_concurrency.lockutils [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] Acquiring lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.924850] env[61857]: DEBUG oslo_concurrency.lockutils [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] Acquired lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.924850] env[61857]: DEBUG nova.network.neutron [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Refreshing network info cache for port 25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 877.137948] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e0189f-c496-b91f-e955-ce4ff7059712, 'name': SearchDatastore_Task, 'duration_secs': 0.012069} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.138250] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.138511] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] d9d05629-79ef-45c6-ac54-b3bc18d306f4/d9d05629-79ef-45c6-ac54-b3bc18d306f4.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 877.138783] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ef5d810-ecdd-448b-a618-5c89badade6e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.147317] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 877.147317] env[61857]: value = "task-951139" [ 877.147317] env[61857]: _type = "Task" [ 877.147317] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.158017] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951139, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.171263] env[61857]: DEBUG nova.compute.manager [req-41ec24c3-6cb7-4f70-8a92-379cb4ecced3 req-e8aa989e-32ee-4aec-9762-22bac29d53ad service nova] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Received event network-changed-6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 877.171474] env[61857]: DEBUG nova.compute.manager [req-41ec24c3-6cb7-4f70-8a92-379cb4ecced3 req-e8aa989e-32ee-4aec-9762-22bac29d53ad service nova] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Refreshing instance network info cache due to event network-changed-6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 877.171809] env[61857]: DEBUG oslo_concurrency.lockutils [req-41ec24c3-6cb7-4f70-8a92-379cb4ecced3 req-e8aa989e-32ee-4aec-9762-22bac29d53ad service nova] Acquiring lock "refresh_cache-0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.172078] env[61857]: DEBUG oslo_concurrency.lockutils [req-41ec24c3-6cb7-4f70-8a92-379cb4ecced3 req-e8aa989e-32ee-4aec-9762-22bac29d53ad service nova] Acquired lock "refresh_cache-0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.172161] env[61857]: DEBUG nova.network.neutron [req-41ec24c3-6cb7-4f70-8a92-379cb4ecced3 req-e8aa989e-32ee-4aec-9762-22bac29d53ad service nova] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Refreshing network info cache for port 6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 877.252847] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951138, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.283501] env[61857]: DEBUG nova.network.neutron [-] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.354386] env[61857]: DEBUG nova.scheduler.client.report [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.666628] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951139, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.757738] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951138, 'name': CreateVM_Task, 'duration_secs': 0.554796} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.757897] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 877.758679] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.758912] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.759332] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.759644] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32b6a2ce-5d5f-433e-b3c0-c68d1a6cfed9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.768241] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for the task: (returnval){ [ 877.768241] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521c6bef-2c71-9991-3192-0980c42ec0de" [ 877.768241] env[61857]: _type = "Task" [ 877.768241] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.780446] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521c6bef-2c71-9991-3192-0980c42ec0de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.786029] env[61857]: INFO nova.compute.manager [-] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Took 1.33 seconds to deallocate network for instance. [ 877.828429] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b3ba2a59-f6d6-4bcd-8e5f-9fb5c81feca4 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquiring lock "interface-6957dc98-8c60-4fdd-83dd-be2f13825c6d-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.828722] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b3ba2a59-f6d6-4bcd-8e5f-9fb5c81feca4 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "interface-6957dc98-8c60-4fdd-83dd-be2f13825c6d-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.829060] env[61857]: DEBUG nova.objects.instance [None req-b3ba2a59-f6d6-4bcd-8e5f-9fb5c81feca4 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lazy-loading 'flavor' on Instance uuid 6957dc98-8c60-4fdd-83dd-be2f13825c6d {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.860354] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.066s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.863167] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.448s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.864707] env[61857]: INFO nova.compute.claims [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.885844] env[61857]: DEBUG nova.network.neutron [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updated VIF entry in instance network info cache for port 25a85668-66ff-4cf6-8cef-68b4d1a2e828. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 877.886295] env[61857]: DEBUG nova.network.neutron [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updating instance_info_cache with network_info: [{"id": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "address": "fa:16:3e:40:07:f3", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25a85668-66", "ovs_interfaceid": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.902238] env[61857]: INFO nova.scheduler.client.report [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Deleted allocations for instance ad89cb5f-44d0-45c9-92a5-7f7ff3138e21 [ 877.905697] env[61857]: DEBUG nova.network.neutron [req-41ec24c3-6cb7-4f70-8a92-379cb4ecced3 req-e8aa989e-32ee-4aec-9762-22bac29d53ad service nova] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Updated VIF entry in instance network info cache for port 6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 877.906659] env[61857]: DEBUG nova.network.neutron [req-41ec24c3-6cb7-4f70-8a92-379cb4ecced3 req-e8aa989e-32ee-4aec-9762-22bac29d53ad service nova] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Updating instance_info_cache with network_info: [{"id": "6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2", "address": "fa:16:3e:3e:cf:79", "network": {"id": "2b3f7088-e451-4d24-bda6-faaf0cac6afa", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-851060206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fb60d26b96c453b8eb4aee57b040d7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a874c214-8cdf-4a41-a718-84262b2a28d8", "external-id": "cl2-zone-726", "segmentation_id": 726, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6dc5c15d-c7", "ovs_interfaceid": "6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.022717] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.023100] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.023233] env[61857]: DEBUG nova.network.neutron [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 878.163244] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951139, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677099} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.163537] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] d9d05629-79ef-45c6-ac54-b3bc18d306f4/d9d05629-79ef-45c6-ac54-b3bc18d306f4.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 878.163761] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.164036] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9a482de-c76f-447b-8e72-d01b2ba701d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.171644] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 878.171644] env[61857]: value = "task-951140" [ 878.171644] env[61857]: _type = "Task" [ 878.171644] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.180158] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951140, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.279281] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521c6bef-2c71-9991-3192-0980c42ec0de, 'name': SearchDatastore_Task, 'duration_secs': 0.03224} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.279606] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.279846] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.280105] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.280284] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.280506] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.280781] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d891475-2270-46c7-bf6e-ff7f812c73ed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.293409] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.293706] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.293940] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 878.294742] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ad6ac84-3e1f-44a0-b273-ca3a367c99d5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.300442] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for the task: (returnval){ [ 878.300442] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]527411b8-c4cf-fc56-e5e7-54000f100c76" [ 878.300442] env[61857]: _type = "Task" [ 878.300442] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.309108] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]527411b8-c4cf-fc56-e5e7-54000f100c76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.333725] env[61857]: DEBUG nova.objects.instance [None req-b3ba2a59-f6d6-4bcd-8e5f-9fb5c81feca4 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lazy-loading 'pci_requests' on Instance uuid 6957dc98-8c60-4fdd-83dd-be2f13825c6d {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.389220] env[61857]: DEBUG oslo_concurrency.lockutils [req-a268fda4-9610-4e32-8d4f-61ac5837cc41 req-9e43d35e-bfde-463d-b844-a94ed455406c service nova] Releasing lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.412038] env[61857]: DEBUG oslo_concurrency.lockutils [req-41ec24c3-6cb7-4f70-8a92-379cb4ecced3 req-e8aa989e-32ee-4aec-9762-22bac29d53ad service nova] Releasing lock "refresh_cache-0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.417219] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1f4d7c44-c249-4fe3-8212-9218425e169c tempest-ServersTestMultiNic-1909134065 tempest-ServersTestMultiNic-1909134065-project-member] Lock "ad89cb5f-44d0-45c9-92a5-7f7ff3138e21" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.211s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.682402] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951140, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.182354} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.682807] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.683682] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d45d1ac-66bc-439f-91ab-04464c55cbc2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.707764] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] d9d05629-79ef-45c6-ac54-b3bc18d306f4/d9d05629-79ef-45c6-ac54-b3bc18d306f4.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.708228] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80578ba7-a628-47c6-9f9c-a3e20068b2ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.729052] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 878.729052] env[61857]: value = "task-951141" [ 878.729052] env[61857]: _type = "Task" [ 878.729052] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.738843] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951141, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.811128] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]527411b8-c4cf-fc56-e5e7-54000f100c76, 'name': SearchDatastore_Task, 'duration_secs': 0.017055} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.811944] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e55f416-d832-4ca0-aea4-a64ecf8e107b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.817840] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for the task: (returnval){ [ 878.817840] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5261a6e4-1d71-395e-0633-e81cb9bf5795" [ 878.817840] env[61857]: _type = "Task" [ 878.817840] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.827703] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5261a6e4-1d71-395e-0633-e81cb9bf5795, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.836728] env[61857]: DEBUG nova.objects.base [None req-b3ba2a59-f6d6-4bcd-8e5f-9fb5c81feca4 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Object Instance<6957dc98-8c60-4fdd-83dd-be2f13825c6d> lazy-loaded attributes: flavor,pci_requests {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 878.837014] env[61857]: DEBUG nova.network.neutron [None req-b3ba2a59-f6d6-4bcd-8e5f-9fb5c81feca4 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 878.956144] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b3ba2a59-f6d6-4bcd-8e5f-9fb5c81feca4 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "interface-6957dc98-8c60-4fdd-83dd-be2f13825c6d-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.127s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.958838] env[61857]: INFO nova.network.neutron [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Port 78436af1-f52d-459b-8460-184ce130e879 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 878.959323] env[61857]: DEBUG nova.network.neutron [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Updating instance_info_cache with network_info: [{"id": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "address": "fa:16:3e:88:8c:fd", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea2b08a8-d6", "ovs_interfaceid": "ea2b08a8-d692-400f-a7a5-124e3a5c14de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.202324] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "16b447b5-426a-4478-9d44-ae32b41dee50" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.202575] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.230128] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.230414] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.230628] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.230817] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.230989] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.237207] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbe2720-c0b5-4bd5-b476-2ab08bd1eedf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.240161] env[61857]: INFO nova.compute.manager [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Terminating instance [ 879.242363] env[61857]: DEBUG nova.compute.manager [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 879.242632] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 879.243980] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d0b57b-f4cd-426a-b2b0-b3eda3762a16 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.251595] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951141, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.252900] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196ab16e-eee5-45ad-b2af-a9e988e26fa0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.257661] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 879.258197] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f067c214-fb1d-4ca6-a0e5-9ba157f68b90 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.287040] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23954849-1ff4-4105-b2cf-d0aecf86ca86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.289726] env[61857]: DEBUG oslo_vmware.api [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 879.289726] env[61857]: value = "task-951142" [ 879.289726] env[61857]: _type = "Task" [ 879.289726] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.296459] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7739ee-209c-4e80-a090-46705b50d361 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.304312] env[61857]: DEBUG oslo_vmware.api [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.314141] env[61857]: DEBUG nova.compute.provider_tree [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.328140] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5261a6e4-1d71-395e-0633-e81cb9bf5795, 'name': SearchDatastore_Task, 'duration_secs': 0.035548} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.328375] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.328628] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc/0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 879.328900] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c95430ce-d8e1-4a36-aee2-e9e0dd965eb6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.335229] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for the task: (returnval){ [ 879.335229] env[61857]: value = "task-951143" [ 879.335229] env[61857]: _type = "Task" [ 879.335229] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.342993] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951143, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.462868] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-96ff157c-4471-4a12-ad99-0aafd6c2dce5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.706752] env[61857]: DEBUG nova.compute.utils [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.743961] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951141, 'name': ReconfigVM_Task, 'duration_secs': 0.909018} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.744297] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Reconfigured VM instance instance-0000004d to attach disk [datastore1] d9d05629-79ef-45c6-ac54-b3bc18d306f4/d9d05629-79ef-45c6-ac54-b3bc18d306f4.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.744964] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f381a7f5-cd1a-4c63-93d9-9ffedae72e32 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.754324] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 879.754324] env[61857]: value = "task-951144" [ 879.754324] env[61857]: _type = "Task" [ 879.754324] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.765063] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951144, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.801110] env[61857]: DEBUG oslo_vmware.api [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951142, 'name': PowerOffVM_Task, 'duration_secs': 0.355246} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.801434] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 879.801628] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 879.801911] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68679172-fd8e-4064-9bb4-14b88ca26528 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.817149] env[61857]: DEBUG nova.scheduler.client.report [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.849147] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951143, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.899327] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 879.899799] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 879.900294] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleting the datastore file [datastore1] 96ff157c-4471-4a12-ad99-0aafd6c2dce5 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.900576] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4db81dac-240e-4860-be3b-2824f36c5fd3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.910260] env[61857]: DEBUG oslo_vmware.api [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 879.910260] env[61857]: value = "task-951146" [ 879.910260] env[61857]: _type = "Task" [ 879.910260] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.918402] env[61857]: DEBUG oslo_vmware.api [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951146, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.969377] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d2a21aab-1c58-4229-95b6-e51c25419a1b tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-96ff157c-4471-4a12-ad99-0aafd6c2dce5-78436af1-f52d-459b-8460-184ce130e879" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.101s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.209963] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.264988] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951144, 'name': Rename_Task, 'duration_secs': 0.240478} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.265252] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 880.265506] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0dc5c311-1301-4c5b-9b6c-0f8041b42623 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.273621] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 880.273621] env[61857]: value = "task-951147" [ 880.273621] env[61857]: _type = "Task" [ 880.273621] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.281630] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951147, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.323116] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.323831] env[61857]: DEBUG nova.compute.manager [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 880.327481] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.272s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.327776] env[61857]: DEBUG nova.objects.instance [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lazy-loading 'resources' on Instance uuid ece27044-3123-4685-a4ab-30eeb10337d1 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.351826] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951143, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562153} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.352185] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc/0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 880.352477] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.352799] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f3a3b575-969b-46cc-98a7-442691916003 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.362348] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for the task: (returnval){ [ 880.362348] env[61857]: value = "task-951148" [ 880.362348] env[61857]: _type = "Task" [ 880.362348] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.374318] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951148, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.421282] env[61857]: DEBUG oslo_vmware.api [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951146, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187239} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.421565] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.421755] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 880.422026] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 880.422139] env[61857]: INFO nova.compute.manager [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Took 1.18 seconds to destroy the instance on the hypervisor. [ 880.422391] env[61857]: DEBUG oslo.service.loopingcall [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.422612] env[61857]: DEBUG nova.compute.manager [-] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 880.422711] env[61857]: DEBUG nova.network.neutron [-] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 880.551834] env[61857]: DEBUG nova.compute.manager [req-3d5d200a-2fa6-4550-9ce6-98cffbec761e req-1aec6e0c-0cee-48d2-9b37-22b0a7ef35ab service nova] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Received event network-vif-deleted-d1542c4d-1685-447b-a6c4-b56fad0ad098 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.785855] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951147, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.835110] env[61857]: DEBUG nova.compute.utils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.837094] env[61857]: DEBUG nova.compute.manager [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 880.837453] env[61857]: DEBUG nova.network.neutron [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 880.873734] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951148, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075638} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.876449] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.878332] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16b683b-b51d-4788-8b42-4971522ede7a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.904722] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc/0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.908634] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aed6a142-71a2-4288-9c52-688366b59a05 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.926195] env[61857]: DEBUG nova.policy [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1496582f8474b14baf63b215a80e64e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3688a3ca85044527a84ad1a7b201fbdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.954534] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for the task: (returnval){ [ 880.954534] env[61857]: value = "task-951149" [ 880.954534] env[61857]: _type = "Task" [ 880.954534] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.975701] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951149, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.196784] env[61857]: DEBUG nova.compute.manager [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Received event network-vif-deleted-78436af1-f52d-459b-8460-184ce130e879 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.196894] env[61857]: DEBUG nova.compute.manager [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Received event network-changed-25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.197052] env[61857]: DEBUG nova.compute.manager [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Refreshing instance network info cache due to event network-changed-25a85668-66ff-4cf6-8cef-68b4d1a2e828. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 881.197379] env[61857]: DEBUG oslo_concurrency.lockutils [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] Acquiring lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.197475] env[61857]: DEBUG oslo_concurrency.lockutils [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] Acquired lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.197622] env[61857]: DEBUG nova.network.neutron [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Refreshing network info cache for port 25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 881.288220] env[61857]: DEBUG oslo_vmware.api [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951147, 'name': PowerOnVM_Task, 'duration_secs': 0.905925} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.289789] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 881.290014] env[61857]: INFO nova.compute.manager [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Took 12.29 seconds to spawn the instance on the hypervisor. [ 881.290235] env[61857]: DEBUG nova.compute.manager [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.291877] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "16b447b5-426a-4478-9d44-ae32b41dee50" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.291877] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.291877] env[61857]: INFO nova.compute.manager [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Attaching volume 53af836e-3d73-4b7e-86c0-3d1641d68f9b to /dev/sdb [ 881.293767] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17408454-98c2-4808-9995-5d23ff711b78 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.340726] env[61857]: DEBUG nova.compute.manager [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 881.357154] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff30ffac-8357-4385-aeab-b6d1c1d0615a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.368622] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bb7975-7142-4ab9-b65c-6c1355fa52fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.380958] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d0890ce-35f4-4e74-9f10-ba7a4a7cf067 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.388411] env[61857]: DEBUG nova.virt.block_device [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Updating existing volume attachment record: 412a2812-4a42-4664-b075-80e9cccfc8d3 {{(pid=61857) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 881.397185] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a074cb30-a359-481b-a134-f419976cb908 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.431857] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2805af7f-b20f-475c-8de6-627d50e97591 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.445759] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57bbb3d-c689-4106-bf3f-106b054914f2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.462318] env[61857]: DEBUG nova.compute.provider_tree [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.473589] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.530727] env[61857]: DEBUG nova.network.neutron [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Successfully created port: 11fe2d50-29b2-4bd7-90ba-b5d587f87ded {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.576447] env[61857]: DEBUG nova.network.neutron [-] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.663642] env[61857]: DEBUG oslo_concurrency.lockutils [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "30d905ed-831d-44ae-807c-062de9a7e9fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.663924] env[61857]: DEBUG oslo_concurrency.lockutils [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.664162] env[61857]: DEBUG oslo_concurrency.lockutils [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "30d905ed-831d-44ae-807c-062de9a7e9fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.664367] env[61857]: DEBUG oslo_concurrency.lockutils [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.664541] env[61857]: DEBUG oslo_concurrency.lockutils [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.666763] env[61857]: INFO nova.compute.manager [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Terminating instance [ 881.670564] env[61857]: DEBUG nova.compute.manager [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.670773] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 881.671638] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877332d2-9181-4031-92f5-31105c1065e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.683385] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 881.686593] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f8eed558-243e-48c7-bf07-12cf8ed79d4f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.702894] env[61857]: DEBUG oslo_vmware.api [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 881.702894] env[61857]: value = "task-951153" [ 881.702894] env[61857]: _type = "Task" [ 881.702894] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.708867] env[61857]: DEBUG oslo_vmware.api [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.824472] env[61857]: INFO nova.compute.manager [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Took 42.77 seconds to build instance. [ 881.941866] env[61857]: DEBUG oslo_concurrency.lockutils [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquiring lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.942239] env[61857]: DEBUG oslo_concurrency.lockutils [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.942500] env[61857]: DEBUG oslo_concurrency.lockutils [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquiring lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.942733] env[61857]: DEBUG oslo_concurrency.lockutils [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.942949] env[61857]: DEBUG oslo_concurrency.lockutils [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.945210] env[61857]: INFO nova.compute.manager [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Terminating instance [ 881.950103] env[61857]: DEBUG nova.compute.manager [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.950274] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 881.951528] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d196d67c-2642-4540-a283-781968e58fbf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.962087] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 881.967593] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e87c775d-95e5-4096-b29e-6a61e8d42ba2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.972541] env[61857]: DEBUG nova.scheduler.client.report [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.984765] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951149, 'name': ReconfigVM_Task, 'duration_secs': 0.702834} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.986515] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc/0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.987430] env[61857]: DEBUG oslo_vmware.api [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for the task: (returnval){ [ 881.987430] env[61857]: value = "task-951154" [ 881.987430] env[61857]: _type = "Task" [ 881.987430] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.987761] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-915380b0-728f-4c79-930b-c799cc159f90 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.001990] env[61857]: DEBUG oslo_vmware.api [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.003441] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for the task: (returnval){ [ 882.003441] env[61857]: value = "task-951155" [ 882.003441] env[61857]: _type = "Task" [ 882.003441] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.013104] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951155, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.079816] env[61857]: INFO nova.compute.manager [-] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Took 1.66 seconds to deallocate network for instance. [ 882.116243] env[61857]: DEBUG nova.network.neutron [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updated VIF entry in instance network info cache for port 25a85668-66ff-4cf6-8cef-68b4d1a2e828. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 882.116685] env[61857]: DEBUG nova.network.neutron [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updating instance_info_cache with network_info: [{"id": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "address": "fa:16:3e:40:07:f3", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25a85668-66", "ovs_interfaceid": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.208381] env[61857]: DEBUG oslo_vmware.api [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951153, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.328017] env[61857]: DEBUG oslo_concurrency.lockutils [None req-22f425f0-76a5-49fc-8ea8-b183a87421ae tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.382s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.333256] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "18a632be-322f-48a9-becf-51ff3b735ad4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.333256] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "18a632be-322f-48a9-becf-51ff3b735ad4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.354675] env[61857]: DEBUG nova.compute.manager [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 882.391952] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='324239e3f2460c6d190f8b2abd89a189',container_format='bare',created_at=2024-09-18T21:43:18Z,direct_url=,disk_format='vmdk',id=de6910ec-d5ea-4309-9229-4103bb53c96c,min_disk=1,min_ram=0,name='tempest-test-snap-1499925240',owner='3688a3ca85044527a84ad1a7b201fbdd',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-09-18T21:43:35Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.392244] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.392415] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.392610] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.392775] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.392913] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.393142] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.393312] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.393488] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.393660] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.393902] env[61857]: DEBUG nova.virt.hardware [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.394755] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a89100-4639-4b45-ab22-bc568cc3f469 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.404896] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12523ceb-299b-4ba3-8854-141f83832ed1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.483089] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.156s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.485432] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.515s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.487389] env[61857]: INFO nova.compute.claims [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.501532] env[61857]: DEBUG oslo_vmware.api [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951154, 'name': PowerOffVM_Task, 'duration_secs': 0.330355} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.501808] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 882.501987] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 882.502269] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68014159-92ba-488c-a46c-29871cd9a683 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.513125] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951155, 'name': Rename_Task, 'duration_secs': 0.235688} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.514283] env[61857]: INFO nova.scheduler.client.report [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Deleted allocations for instance ece27044-3123-4685-a4ab-30eeb10337d1 [ 882.515205] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 882.515498] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-071ea0a0-9fba-46bf-a411-50be863cc8c4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.528579] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for the task: (returnval){ [ 882.528579] env[61857]: value = "task-951157" [ 882.528579] env[61857]: _type = "Task" [ 882.528579] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.540483] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.587405] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.621564] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 882.621564] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 882.621564] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Deleting the datastore file [datastore1] 6957dc98-8c60-4fdd-83dd-be2f13825c6d {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.621564] env[61857]: DEBUG oslo_concurrency.lockutils [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] Releasing lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.621564] env[61857]: DEBUG nova.compute.manager [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Received event network-changed-7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 882.621564] env[61857]: DEBUG nova.compute.manager [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Refreshing instance network info cache due to event network-changed-7ede4f79-7428-48b0-a811-e8c064d224f0. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 882.621564] env[61857]: DEBUG oslo_concurrency.lockutils [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] Acquiring lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.621564] env[61857]: DEBUG oslo_concurrency.lockutils [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] Acquired lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.621564] env[61857]: DEBUG nova.network.neutron [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Refreshing network info cache for port 7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 882.622816] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d526627-d9d4-4c66-9d9d-8b4dbd9ae000 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.630868] env[61857]: DEBUG oslo_vmware.api [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for the task: (returnval){ [ 882.630868] env[61857]: value = "task-951158" [ 882.630868] env[61857]: _type = "Task" [ 882.630868] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.640237] env[61857]: DEBUG oslo_vmware.api [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951158, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.708561] env[61857]: DEBUG oslo_vmware.api [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951153, 'name': PowerOffVM_Task, 'duration_secs': 0.62086} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.708984] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 882.709226] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 882.709619] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79e0fcf6-042a-4479-a0a4-851a68001706 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.803904] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 882.803904] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 882.804385] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleting the datastore file [datastore2] 30d905ed-831d-44ae-807c-062de9a7e9fb {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.804385] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02d8e092-7535-4ace-ad96-4ca16e790ffa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.813650] env[61857]: DEBUG oslo_vmware.api [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 882.813650] env[61857]: value = "task-951160" [ 882.813650] env[61857]: _type = "Task" [ 882.813650] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.821730] env[61857]: DEBUG oslo_vmware.api [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951160, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.824808] env[61857]: DEBUG nova.compute.manager [req-492dee75-09d5-47fb-b4c0-8595ff5e0857 req-bf229b06-3a87-4aee-97a5-2678f9c427ac service nova] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Received event network-vif-deleted-ea2b08a8-d692-400f-a7a5-124e3a5c14de {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 882.837320] env[61857]: DEBUG nova.compute.manager [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.028053] env[61857]: DEBUG oslo_concurrency.lockutils [None req-67e8f493-ea17-4a9a-a317-95a9a360708e tempest-ServerShowV257Test-337184745 tempest-ServerShowV257Test-337184745-project-member] Lock "ece27044-3123-4685-a4ab-30eeb10337d1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.701s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.041580] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951157, 'name': PowerOnVM_Task} progress is 79%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.146758] env[61857]: DEBUG oslo_vmware.api [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Task: {'id': task-951158, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.356742} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.147316] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.147554] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 883.147760] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 883.148200] env[61857]: INFO nova.compute.manager [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Took 1.20 seconds to destroy the instance on the hypervisor. [ 883.148489] env[61857]: DEBUG oslo.service.loopingcall [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.148700] env[61857]: DEBUG nova.compute.manager [-] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.149360] env[61857]: DEBUG nova.network.neutron [-] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 883.324826] env[61857]: DEBUG oslo_vmware.api [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951160, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.396987} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.325159] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.325426] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 883.325541] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 883.325726] env[61857]: INFO nova.compute.manager [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Took 1.65 seconds to destroy the instance on the hypervisor. [ 883.325977] env[61857]: DEBUG oslo.service.loopingcall [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.326201] env[61857]: DEBUG nova.compute.manager [-] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.326296] env[61857]: DEBUG nova.network.neutron [-] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 883.345833] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "ffa63de7-dd32-4908-92c5-755b3c0799ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.346123] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.346346] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "ffa63de7-dd32-4908-92c5-755b3c0799ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.346542] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.346719] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.351223] env[61857]: INFO nova.compute.manager [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Terminating instance [ 883.354797] env[61857]: DEBUG nova.compute.manager [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 883.354890] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 883.356289] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe24f156-8c89-4c34-9c06-1eb0e41ac260 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.365463] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 883.367618] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe50ff15-55b1-4621-85c2-a9aeac8c9f68 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.370550] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.377181] env[61857]: DEBUG oslo_vmware.api [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 883.377181] env[61857]: value = "task-951161" [ 883.377181] env[61857]: _type = "Task" [ 883.377181] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.386642] env[61857]: DEBUG oslo_vmware.api [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951161, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.546261] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951157, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.634013] env[61857]: DEBUG nova.compute.manager [req-035a0e37-570e-47d1-a01c-18fb507941f5 req-e3d9a7ae-d6d7-4230-80cb-f541d20a2842 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Received event network-changed-7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.634350] env[61857]: DEBUG nova.compute.manager [req-035a0e37-570e-47d1-a01c-18fb507941f5 req-e3d9a7ae-d6d7-4230-80cb-f541d20a2842 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Refreshing instance network info cache due to event network-changed-7ede4f79-7428-48b0-a811-e8c064d224f0. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 883.634607] env[61857]: DEBUG oslo_concurrency.lockutils [req-035a0e37-570e-47d1-a01c-18fb507941f5 req-e3d9a7ae-d6d7-4230-80cb-f541d20a2842 service nova] Acquiring lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.731687] env[61857]: DEBUG nova.network.neutron [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Updated VIF entry in instance network info cache for port 7ede4f79-7428-48b0-a811-e8c064d224f0. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 883.733480] env[61857]: DEBUG nova.network.neutron [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Updating instance_info_cache with network_info: [{"id": "7ede4f79-7428-48b0-a811-e8c064d224f0", "address": "fa:16:3e:9e:cc:3a", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ede4f79-74", "ovs_interfaceid": "7ede4f79-7428-48b0-a811-e8c064d224f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.898511] env[61857]: DEBUG oslo_vmware.api [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951161, 'name': PowerOffVM_Task, 'duration_secs': 0.301025} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.899813] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 883.899813] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 883.899813] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f7a79544-7095-4236-b743-59e909a17de6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.990607] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 883.990607] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 883.990607] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Deleting the datastore file [datastore2] ffa63de7-dd32-4908-92c5-755b3c0799ef {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.990607] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96f8d68b-b509-4c17-81fd-7c70f76715b9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.998344] env[61857]: DEBUG oslo_vmware.api [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 883.998344] env[61857]: value = "task-951164" [ 883.998344] env[61857]: _type = "Task" [ 883.998344] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.013990] env[61857]: DEBUG oslo_vmware.api [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.021664] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a30f13-50e4-4db5-a31d-87f0a4c70465 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.025875] env[61857]: DEBUG nova.network.neutron [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Successfully updated port: 11fe2d50-29b2-4bd7-90ba-b5d587f87ded {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 884.030758] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5ad0a6-e14b-481b-a2d4-7e9dbe659720 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.047817] env[61857]: DEBUG oslo_vmware.api [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951157, 'name': PowerOnVM_Task, 'duration_secs': 1.289317} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.074747] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 884.075034] env[61857]: INFO nova.compute.manager [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Took 9.92 seconds to spawn the instance on the hypervisor. [ 884.075303] env[61857]: DEBUG nova.compute.manager [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.076618] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aba4b9f-4667-4578-86c5-6867adbc75d8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.079771] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2842242-fc1f-4d70-8b2e-133a4a5e1ea8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.090053] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09898560-b3a3-47f7-b3ed-e78e58af2b4c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.108688] env[61857]: DEBUG nova.compute.provider_tree [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.236448] env[61857]: DEBUG oslo_concurrency.lockutils [req-12f78f1a-5a0b-4b0a-b894-24754bb14c0c req-ee477cef-9047-4f60-9807-8dccdac45220 service nova] Releasing lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.236914] env[61857]: DEBUG oslo_concurrency.lockutils [req-035a0e37-570e-47d1-a01c-18fb507941f5 req-e3d9a7ae-d6d7-4230-80cb-f541d20a2842 service nova] Acquired lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.237126] env[61857]: DEBUG nova.network.neutron [req-035a0e37-570e-47d1-a01c-18fb507941f5 req-e3d9a7ae-d6d7-4230-80cb-f541d20a2842 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Refreshing network info cache for port 7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 884.261554] env[61857]: DEBUG nova.compute.manager [req-f3a7a04e-8a3b-4121-9a88-8998979c6a9a req-43d1f135-9cf9-49c1-9e73-29fa77eabd2b service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Received event network-vif-plugged-11fe2d50-29b2-4bd7-90ba-b5d587f87ded {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.261792] env[61857]: DEBUG oslo_concurrency.lockutils [req-f3a7a04e-8a3b-4121-9a88-8998979c6a9a req-43d1f135-9cf9-49c1-9e73-29fa77eabd2b service nova] Acquiring lock "33cb5bbc-926d-42ee-b483-8d1e24707e40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.262024] env[61857]: DEBUG oslo_concurrency.lockutils [req-f3a7a04e-8a3b-4121-9a88-8998979c6a9a req-43d1f135-9cf9-49c1-9e73-29fa77eabd2b service nova] Lock "33cb5bbc-926d-42ee-b483-8d1e24707e40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.262292] env[61857]: DEBUG oslo_concurrency.lockutils [req-f3a7a04e-8a3b-4121-9a88-8998979c6a9a req-43d1f135-9cf9-49c1-9e73-29fa77eabd2b service nova] Lock "33cb5bbc-926d-42ee-b483-8d1e24707e40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.262471] env[61857]: DEBUG nova.compute.manager [req-f3a7a04e-8a3b-4121-9a88-8998979c6a9a req-43d1f135-9cf9-49c1-9e73-29fa77eabd2b service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] No waiting events found dispatching network-vif-plugged-11fe2d50-29b2-4bd7-90ba-b5d587f87ded {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 884.262555] env[61857]: WARNING nova.compute.manager [req-f3a7a04e-8a3b-4121-9a88-8998979c6a9a req-43d1f135-9cf9-49c1-9e73-29fa77eabd2b service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Received unexpected event network-vif-plugged-11fe2d50-29b2-4bd7-90ba-b5d587f87ded for instance with vm_state building and task_state spawning. [ 884.520289] env[61857]: DEBUG oslo_vmware.api [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253899} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.520657] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.521569] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 884.525184] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 884.525508] env[61857]: INFO nova.compute.manager [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Took 1.17 seconds to destroy the instance on the hypervisor. [ 884.525858] env[61857]: DEBUG oslo.service.loopingcall [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.526144] env[61857]: DEBUG nova.compute.manager [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 884.526818] env[61857]: DEBUG nova.network.neutron [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 884.530113] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "refresh_cache-33cb5bbc-926d-42ee-b483-8d1e24707e40" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.530304] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "refresh_cache-33cb5bbc-926d-42ee-b483-8d1e24707e40" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.530534] env[61857]: DEBUG nova.network.neutron [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 884.607082] env[61857]: INFO nova.compute.manager [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Took 36.21 seconds to build instance. [ 884.611192] env[61857]: DEBUG nova.scheduler.client.report [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.759782] env[61857]: DEBUG nova.network.neutron [-] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.783142] env[61857]: DEBUG nova.network.neutron [-] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.109102] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5c5c4eea-d45c-4830-8033-eaed2990f008 tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.850s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.123123] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.127020] env[61857]: DEBUG nova.compute.manager [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.127020] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.948s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.127292] env[61857]: DEBUG nova.objects.instance [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lazy-loading 'resources' on Instance uuid d29f6e34-861d-4cbf-8cbd-ab193a55220b {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.138433] env[61857]: DEBUG nova.network.neutron [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 885.262250] env[61857]: INFO nova.compute.manager [-] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Took 2.11 seconds to deallocate network for instance. [ 885.286487] env[61857]: INFO nova.compute.manager [-] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Took 1.96 seconds to deallocate network for instance. [ 885.364671] env[61857]: DEBUG nova.network.neutron [req-035a0e37-570e-47d1-a01c-18fb507941f5 req-e3d9a7ae-d6d7-4230-80cb-f541d20a2842 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Updated VIF entry in instance network info cache for port 7ede4f79-7428-48b0-a811-e8c064d224f0. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 885.365238] env[61857]: DEBUG nova.network.neutron [req-035a0e37-570e-47d1-a01c-18fb507941f5 req-e3d9a7ae-d6d7-4230-80cb-f541d20a2842 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Updating instance_info_cache with network_info: [{"id": "7ede4f79-7428-48b0-a811-e8c064d224f0", "address": "fa:16:3e:9e:cc:3a", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ede4f79-74", "ovs_interfaceid": "7ede4f79-7428-48b0-a811-e8c064d224f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.447582] env[61857]: DEBUG nova.network.neutron [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Updating instance_info_cache with network_info: [{"id": "11fe2d50-29b2-4bd7-90ba-b5d587f87ded", "address": "fa:16:3e:30:ce:c2", "network": {"id": "e7b01bbf-aac5-42cb-b166-da2a3a70a175", "bridge": "br-int", "label": "tempest-ImagesTestJSON-229290731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3688a3ca85044527a84ad1a7b201fbdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11fe2d50-29", "ovs_interfaceid": "11fe2d50-29b2-4bd7-90ba-b5d587f87ded", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.604738] env[61857]: DEBUG nova.network.neutron [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.637351] env[61857]: DEBUG nova.compute.utils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.639759] env[61857]: DEBUG nova.compute.manager [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.640015] env[61857]: DEBUG nova.network.neutron [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 885.642500] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquiring lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.642756] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.642934] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquiring lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.643152] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.643395] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.645921] env[61857]: INFO nova.compute.manager [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Terminating instance [ 885.650317] env[61857]: DEBUG nova.compute.manager [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 885.650542] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 885.651429] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24ee60a-8fd3-4ff6-9081-e5ebd3d8e4d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.668938] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 885.669347] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef99bf9e-75a3-4a2b-a4c5-738f53e4048c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.683630] env[61857]: DEBUG oslo_vmware.api [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for the task: (returnval){ [ 885.683630] env[61857]: value = "task-951165" [ 885.683630] env[61857]: _type = "Task" [ 885.683630] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.691816] env[61857]: DEBUG nova.compute.manager [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Received event network-changed-11fe2d50-29b2-4bd7-90ba-b5d587f87ded {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 885.691816] env[61857]: DEBUG nova.compute.manager [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Refreshing instance network info cache due to event network-changed-11fe2d50-29b2-4bd7-90ba-b5d587f87ded. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 885.692018] env[61857]: DEBUG oslo_concurrency.lockutils [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] Acquiring lock "refresh_cache-33cb5bbc-926d-42ee-b483-8d1e24707e40" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.704631] env[61857]: DEBUG oslo_vmware.api [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951165, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.712671] env[61857]: DEBUG nova.policy [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74f5499cbed541deaf1905605e10158f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62bd314daba846cd9ed3dc994dd1034b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.769552] env[61857]: DEBUG oslo_concurrency.lockutils [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.791482] env[61857]: DEBUG oslo_concurrency.lockutils [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.871673] env[61857]: DEBUG oslo_concurrency.lockutils [req-035a0e37-570e-47d1-a01c-18fb507941f5 req-e3d9a7ae-d6d7-4230-80cb-f541d20a2842 service nova] Releasing lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.951823] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "refresh_cache-33cb5bbc-926d-42ee-b483-8d1e24707e40" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.952224] env[61857]: DEBUG nova.compute.manager [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Instance network_info: |[{"id": "11fe2d50-29b2-4bd7-90ba-b5d587f87ded", "address": "fa:16:3e:30:ce:c2", "network": {"id": "e7b01bbf-aac5-42cb-b166-da2a3a70a175", "bridge": "br-int", "label": "tempest-ImagesTestJSON-229290731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3688a3ca85044527a84ad1a7b201fbdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11fe2d50-29", "ovs_interfaceid": "11fe2d50-29b2-4bd7-90ba-b5d587f87ded", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 885.952573] env[61857]: DEBUG oslo_concurrency.lockutils [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] Acquired lock "refresh_cache-33cb5bbc-926d-42ee-b483-8d1e24707e40" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.952892] env[61857]: DEBUG nova.network.neutron [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Refreshing network info cache for port 11fe2d50-29b2-4bd7-90ba-b5d587f87ded {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 885.954456] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:ce:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4576b9d4-535c-40aa-b078-246f671f216e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '11fe2d50-29b2-4bd7-90ba-b5d587f87ded', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.970249] env[61857]: DEBUG oslo.service.loopingcall [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.977373] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Volume attach. Driver type: vmdk {{(pid=61857) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 885.977373] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214149', 'volume_id': '53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'name': 'volume-53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '16b447b5-426a-4478-9d44-ae32b41dee50', 'attached_at': '', 'detached_at': '', 'volume_id': '53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'serial': '53af836e-3d73-4b7e-86c0-3d1641d68f9b'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 885.978284] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 885.979080] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587b6345-8413-4d5f-aedb-734cc580e4a8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.983018] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d915099f-0014-4d83-8801-f4b0097243ee {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.024838] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa6223a-cfbf-414a-9e4a-061af407e47a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.030812] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.030812] env[61857]: value = "task-951166" [ 886.030812] env[61857]: _type = "Task" [ 886.030812] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.059951] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] volume-53af836e-3d73-4b7e-86c0-3d1641d68f9b/volume-53af836e-3d73-4b7e-86c0-3d1641d68f9b.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.062933] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-025fa425-254e-4550-82c8-3279e08906f6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.079307] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951166, 'name': CreateVM_Task} progress is 15%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.086106] env[61857]: DEBUG oslo_vmware.api [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 886.086106] env[61857]: value = "task-951167" [ 886.086106] env[61857]: _type = "Task" [ 886.086106] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.109031] env[61857]: INFO nova.compute.manager [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Took 1.58 seconds to deallocate network for instance. [ 886.109405] env[61857]: DEBUG oslo_vmware.api [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951167, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.142796] env[61857]: DEBUG nova.compute.manager [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.189899] env[61857]: DEBUG nova.network.neutron [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Successfully created port: 4ba4d151-0e84-4f6d-aee1-6843bae4fe64 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.201753] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a018edfa-b1bc-48fa-830b-1bc6c3a8eb99 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.212904] env[61857]: DEBUG oslo_vmware.api [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951165, 'name': PowerOffVM_Task, 'duration_secs': 0.354094} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.213897] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa59ee7-70e6-42f9-a981-d7307124216e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.217271] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 886.217474] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 886.217751] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-125178a3-c5b3-414d-8a1b-c733550ace1a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.252860] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2cd286e-6b51-4e7c-b037-1a5beb9122d5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.261520] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232d8748-bf22-4b80-ab1a-f299272aee32 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.277814] env[61857]: DEBUG nova.compute.provider_tree [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.319476] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 886.319715] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 886.320092] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Deleting the datastore file [datastore1] 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.320274] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c994c913-a1f6-4488-893d-c0807804809f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.330553] env[61857]: DEBUG oslo_vmware.api [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for the task: (returnval){ [ 886.330553] env[61857]: value = "task-951169" [ 886.330553] env[61857]: _type = "Task" [ 886.330553] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.340952] env[61857]: DEBUG oslo_vmware.api [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951169, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.543428] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951166, 'name': CreateVM_Task, 'duration_secs': 0.460252} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.543616] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 886.544579] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.544864] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "[datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.545373] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 886.546270] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30263349-5675-471b-8327-bf5c18b611cf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.554062] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 886.554062] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52bc192e-a911-8469-625b-a79bf10358d9" [ 886.554062] env[61857]: _type = "Task" [ 886.554062] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.575291] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "[datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.575291] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Processing image de6910ec-d5ea-4309-9229-4103bb53c96c {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.575291] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c/de6910ec-d5ea-4309-9229-4103bb53c96c.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.575291] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "[datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c/de6910ec-d5ea-4309-9229-4103bb53c96c.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.575291] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.575291] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cc73ddf2-fa5e-4e4f-a05e-781f486e745a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.591351] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.591586] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 886.592409] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-191bab97-c6d7-4039-8006-cd8cfc8a06b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.600297] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 886.600297] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5219a7e1-b409-74ff-ec3b-a3910ea8f6ed" [ 886.600297] env[61857]: _type = "Task" [ 886.600297] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.603743] env[61857]: DEBUG oslo_vmware.api [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951167, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.617758] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Preparing fetch location {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 886.618047] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Fetch image to [datastore1] OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f/OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f.vmdk {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 886.618247] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Downloading stream optimized image de6910ec-d5ea-4309-9229-4103bb53c96c to [datastore1] OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f/OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f.vmdk on the data store datastore1 as vApp {{(pid=61857) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 886.618423] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Downloading image file data de6910ec-d5ea-4309-9229-4103bb53c96c to the ESX as VM named 'OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f' {{(pid=61857) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 886.621515] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.712737] env[61857]: DEBUG oslo_vmware.rw_handles [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 886.712737] env[61857]: value = "resgroup-9" [ 886.712737] env[61857]: _type = "ResourcePool" [ 886.712737] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 886.713073] env[61857]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-79233628-9012-4557-83b1-eec211a2d126 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.749276] env[61857]: DEBUG oslo_vmware.rw_handles [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lease: (returnval){ [ 886.749276] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52efa3b0-0592-e821-7b70-eb4577627aba" [ 886.749276] env[61857]: _type = "HttpNfcLease" [ 886.749276] env[61857]: } obtained for vApp import into resource pool (val){ [ 886.749276] env[61857]: value = "resgroup-9" [ 886.749276] env[61857]: _type = "ResourcePool" [ 886.749276] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 886.749567] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the lease: (returnval){ [ 886.749567] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52efa3b0-0592-e821-7b70-eb4577627aba" [ 886.749567] env[61857]: _type = "HttpNfcLease" [ 886.749567] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 886.765753] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 886.765753] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52efa3b0-0592-e821-7b70-eb4577627aba" [ 886.765753] env[61857]: _type = "HttpNfcLease" [ 886.765753] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 886.779651] env[61857]: DEBUG nova.scheduler.client.report [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.785643] env[61857]: DEBUG nova.compute.manager [req-80e66bfe-1c13-4f1a-b918-fdda89ae91bf req-d8f9f400-e304-47b4-8f44-cd1b7575c760 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Received event network-changed-25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.785643] env[61857]: DEBUG nova.compute.manager [req-80e66bfe-1c13-4f1a-b918-fdda89ae91bf req-d8f9f400-e304-47b4-8f44-cd1b7575c760 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Refreshing instance network info cache due to event network-changed-25a85668-66ff-4cf6-8cef-68b4d1a2e828. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 886.785800] env[61857]: DEBUG oslo_concurrency.lockutils [req-80e66bfe-1c13-4f1a-b918-fdda89ae91bf req-d8f9f400-e304-47b4-8f44-cd1b7575c760 service nova] Acquiring lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.785954] env[61857]: DEBUG oslo_concurrency.lockutils [req-80e66bfe-1c13-4f1a-b918-fdda89ae91bf req-d8f9f400-e304-47b4-8f44-cd1b7575c760 service nova] Acquired lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.786351] env[61857]: DEBUG nova.network.neutron [req-80e66bfe-1c13-4f1a-b918-fdda89ae91bf req-d8f9f400-e304-47b4-8f44-cd1b7575c760 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Refreshing network info cache for port 25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 886.844567] env[61857]: DEBUG oslo_vmware.api [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Task: {'id': task-951169, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.246957} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.844853] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.849027] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 886.849027] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 886.849027] env[61857]: INFO nova.compute.manager [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Took 1.20 seconds to destroy the instance on the hypervisor. [ 886.849027] env[61857]: DEBUG oslo.service.loopingcall [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.849027] env[61857]: DEBUG nova.compute.manager [-] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 886.849027] env[61857]: DEBUG nova.network.neutron [-] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 886.859300] env[61857]: DEBUG nova.network.neutron [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Updated VIF entry in instance network info cache for port 11fe2d50-29b2-4bd7-90ba-b5d587f87ded. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 886.859686] env[61857]: DEBUG nova.network.neutron [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Updating instance_info_cache with network_info: [{"id": "11fe2d50-29b2-4bd7-90ba-b5d587f87ded", "address": "fa:16:3e:30:ce:c2", "network": {"id": "e7b01bbf-aac5-42cb-b166-da2a3a70a175", "bridge": "br-int", "label": "tempest-ImagesTestJSON-229290731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3688a3ca85044527a84ad1a7b201fbdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11fe2d50-29", "ovs_interfaceid": "11fe2d50-29b2-4bd7-90ba-b5d587f87ded", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.101802] env[61857]: DEBUG oslo_vmware.api [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951167, 'name': ReconfigVM_Task, 'duration_secs': 0.533862} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.102237] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Reconfigured VM instance instance-00000042 to attach disk [datastore1] volume-53af836e-3d73-4b7e-86c0-3d1641d68f9b/volume-53af836e-3d73-4b7e-86c0-3d1641d68f9b.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.108190] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81cc519e-6079-42ee-b736-10d520b3875f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.132593] env[61857]: DEBUG oslo_vmware.api [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 887.132593] env[61857]: value = "task-951171" [ 887.132593] env[61857]: _type = "Task" [ 887.132593] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.144146] env[61857]: DEBUG oslo_vmware.api [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951171, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.168666] env[61857]: DEBUG nova.compute.manager [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.207142] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.207142] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.207388] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.208065] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.208065] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.208065] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.208065] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.208304] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.208526] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.208668] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.208845] env[61857]: DEBUG nova.virt.hardware [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.210656] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58ec2c0-e949-489b-96e3-512e00d82528 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.221860] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42276a7a-9f5a-4eed-be95-cb327c718404 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.258892] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.258892] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52efa3b0-0592-e821-7b70-eb4577627aba" [ 887.258892] env[61857]: _type = "HttpNfcLease" [ 887.258892] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 887.289027] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.162s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.294205] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.721s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.294205] env[61857]: DEBUG nova.objects.instance [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lazy-loading 'resources' on Instance uuid 3fb5d24b-1767-43d9-bed5-833121962adb {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.317419] env[61857]: INFO nova.scheduler.client.report [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Deleted allocations for instance d29f6e34-861d-4cbf-8cbd-ab193a55220b [ 887.364016] env[61857]: DEBUG oslo_concurrency.lockutils [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] Releasing lock "refresh_cache-33cb5bbc-926d-42ee-b483-8d1e24707e40" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.364522] env[61857]: DEBUG nova.compute.manager [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Received event network-changed-d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.364838] env[61857]: DEBUG nova.compute.manager [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Refreshing instance network info cache due to event network-changed-d9f80f94-9be0-4d62-863f-7562c6c9e63e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 887.365225] env[61857]: DEBUG oslo_concurrency.lockutils [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] Acquiring lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.365513] env[61857]: DEBUG oslo_concurrency.lockutils [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] Acquired lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.365835] env[61857]: DEBUG nova.network.neutron [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Refreshing network info cache for port d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 887.532115] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "ffa63de7-dd32-4908-92c5-755b3c0799ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.662280] env[61857]: DEBUG oslo_vmware.api [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951171, 'name': ReconfigVM_Task, 'duration_secs': 0.156645} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.664399] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214149', 'volume_id': '53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'name': 'volume-53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '16b447b5-426a-4478-9d44-ae32b41dee50', 'attached_at': '', 'detached_at': '', 'volume_id': '53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'serial': '53af836e-3d73-4b7e-86c0-3d1641d68f9b'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 887.696879] env[61857]: DEBUG nova.network.neutron [-] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.732575] env[61857]: DEBUG nova.compute.manager [req-86c3bde2-1205-40b8-9b76-e20712044b8c req-35f964dc-671b-4556-9ae8-840e19b806e4 service nova] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Received event network-vif-deleted-6dc5c15d-c76e-4d86-9a66-25a93d7e1ab2 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.758275] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.758275] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52efa3b0-0592-e821-7b70-eb4577627aba" [ 887.758275] env[61857]: _type = "HttpNfcLease" [ 887.758275] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 887.825466] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c7247c1a-cae9-427f-a35f-629d21868b07 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "d29f6e34-861d-4cbf-8cbd-ab193a55220b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.801s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.867046] env[61857]: DEBUG nova.network.neutron [req-80e66bfe-1c13-4f1a-b918-fdda89ae91bf req-d8f9f400-e304-47b4-8f44-cd1b7575c760 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updated VIF entry in instance network info cache for port 25a85668-66ff-4cf6-8cef-68b4d1a2e828. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 887.867697] env[61857]: DEBUG nova.network.neutron [req-80e66bfe-1c13-4f1a-b918-fdda89ae91bf req-d8f9f400-e304-47b4-8f44-cd1b7575c760 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updating instance_info_cache with network_info: [{"id": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "address": "fa:16:3e:40:07:f3", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25a85668-66", "ovs_interfaceid": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.919189] env[61857]: DEBUG nova.compute.manager [req-40b82dea-4177-4321-835a-264e79b8bbe9 req-e2e8b841-b920-41fc-9bca-f6cd45b7a702 service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Received event network-vif-plugged-4ba4d151-0e84-4f6d-aee1-6843bae4fe64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.919467] env[61857]: DEBUG oslo_concurrency.lockutils [req-40b82dea-4177-4321-835a-264e79b8bbe9 req-e2e8b841-b920-41fc-9bca-f6cd45b7a702 service nova] Acquiring lock "d979a8a1-a5b5-428e-9b60-7d827337218f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.919689] env[61857]: DEBUG oslo_concurrency.lockutils [req-40b82dea-4177-4321-835a-264e79b8bbe9 req-e2e8b841-b920-41fc-9bca-f6cd45b7a702 service nova] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.919969] env[61857]: DEBUG oslo_concurrency.lockutils [req-40b82dea-4177-4321-835a-264e79b8bbe9 req-e2e8b841-b920-41fc-9bca-f6cd45b7a702 service nova] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.920259] env[61857]: DEBUG nova.compute.manager [req-40b82dea-4177-4321-835a-264e79b8bbe9 req-e2e8b841-b920-41fc-9bca-f6cd45b7a702 service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] No waiting events found dispatching network-vif-plugged-4ba4d151-0e84-4f6d-aee1-6843bae4fe64 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.920458] env[61857]: WARNING nova.compute.manager [req-40b82dea-4177-4321-835a-264e79b8bbe9 req-e2e8b841-b920-41fc-9bca-f6cd45b7a702 service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Received unexpected event network-vif-plugged-4ba4d151-0e84-4f6d-aee1-6843bae4fe64 for instance with vm_state building and task_state spawning. [ 888.201020] env[61857]: INFO nova.compute.manager [-] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Took 1.35 seconds to deallocate network for instance. [ 888.252103] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c69ff7-20d7-46de-9e86-b57279387147 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.271027] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6bea7a3-2319-48b6-952a-3efdad98ad6d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.273573] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 888.273573] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52efa3b0-0592-e821-7b70-eb4577627aba" [ 888.273573] env[61857]: _type = "HttpNfcLease" [ 888.273573] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 888.304472] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35331d3-2463-42ac-b25d-6dd9ba1691bd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.314789] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192f8642-d9c9-47da-b336-d27145178a8e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.330733] env[61857]: DEBUG nova.compute.provider_tree [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.342092] env[61857]: DEBUG nova.network.neutron [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updated VIF entry in instance network info cache for port d9f80f94-9be0-4d62-863f-7562c6c9e63e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 888.342352] env[61857]: DEBUG nova.network.neutron [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updating instance_info_cache with network_info: [{"id": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "address": "fa:16:3e:94:19:6b", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9f80f94-9b", "ovs_interfaceid": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.370622] env[61857]: DEBUG oslo_concurrency.lockutils [req-80e66bfe-1c13-4f1a-b918-fdda89ae91bf req-d8f9f400-e304-47b4-8f44-cd1b7575c760 service nova] Releasing lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.491724] env[61857]: DEBUG nova.network.neutron [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Successfully updated port: 4ba4d151-0e84-4f6d-aee1-6843bae4fe64 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.706647] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.723846] env[61857]: DEBUG nova.objects.instance [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lazy-loading 'flavor' on Instance uuid 16b447b5-426a-4478-9d44-ae32b41dee50 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.759641] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 888.759641] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52efa3b0-0592-e821-7b70-eb4577627aba" [ 888.759641] env[61857]: _type = "HttpNfcLease" [ 888.759641] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 888.833908] env[61857]: DEBUG nova.scheduler.client.report [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.845014] env[61857]: DEBUG oslo_concurrency.lockutils [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] Releasing lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.845306] env[61857]: DEBUG nova.compute.manager [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Received event network-vif-deleted-feebcdf7-ef87-4705-90cf-ebb4415a92fc {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.845528] env[61857]: DEBUG nova.compute.manager [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Received event network-vif-deleted-9d8efcbb-1103-49f6-a52f-219e1682d275 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.845722] env[61857]: DEBUG nova.compute.manager [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Received event network-changed-25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.845889] env[61857]: DEBUG nova.compute.manager [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Refreshing instance network info cache due to event network-changed-25a85668-66ff-4cf6-8cef-68b4d1a2e828. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 888.846113] env[61857]: DEBUG oslo_concurrency.lockutils [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] Acquiring lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.846264] env[61857]: DEBUG oslo_concurrency.lockutils [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] Acquired lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.846432] env[61857]: DEBUG nova.network.neutron [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Refreshing network info cache for port 25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 888.994232] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "ddd4c436-405f-49f7-8c9b-de3b71725f63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.994565] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "ddd4c436-405f-49f7-8c9b-de3b71725f63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.994815] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "ddd4c436-405f-49f7-8c9b-de3b71725f63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.995060] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "ddd4c436-405f-49f7-8c9b-de3b71725f63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.995268] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "ddd4c436-405f-49f7-8c9b-de3b71725f63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.997642] env[61857]: INFO nova.compute.manager [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Terminating instance [ 888.999511] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "refresh_cache-d979a8a1-a5b5-428e-9b60-7d827337218f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.999620] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "refresh_cache-d979a8a1-a5b5-428e-9b60-7d827337218f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.999769] env[61857]: DEBUG nova.network.neutron [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 889.001842] env[61857]: DEBUG nova.compute.manager [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.002093] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 889.003152] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6b9ccb-ae85-4003-b5e6-5b0488c496b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.016354] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 889.016877] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f11f22a0-4f8d-407a-8016-900d3ab0825f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.024246] env[61857]: DEBUG oslo_vmware.api [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 889.024246] env[61857]: value = "task-951172" [ 889.024246] env[61857]: _type = "Task" [ 889.024246] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.034374] env[61857]: DEBUG oslo_vmware.api [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951172, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.056018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "7a9252da-2584-40ed-9d28-ca7341ed5165" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.056331] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "7a9252da-2584-40ed-9d28-ca7341ed5165" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.056557] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "7a9252da-2584-40ed-9d28-ca7341ed5165-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.056751] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "7a9252da-2584-40ed-9d28-ca7341ed5165-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.056981] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "7a9252da-2584-40ed-9d28-ca7341ed5165-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.059473] env[61857]: INFO nova.compute.manager [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Terminating instance [ 889.061451] env[61857]: DEBUG nova.compute.manager [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.061660] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 889.062500] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d8f484-5fcb-4736-8723-4ff3a2f4f835 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.070584] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 889.070855] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd09e16c-15c5-4256-a1a7-f06f4cc6285a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.078416] env[61857]: DEBUG oslo_vmware.api [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 889.078416] env[61857]: value = "task-951173" [ 889.078416] env[61857]: _type = "Task" [ 889.078416] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.087672] env[61857]: DEBUG oslo_vmware.api [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951173, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.230035] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51f4e0af-4f6d-410c-a614-36e6fd3378e2 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.939s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.260942] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 889.260942] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52efa3b0-0592-e821-7b70-eb4577627aba" [ 889.260942] env[61857]: _type = "HttpNfcLease" [ 889.260942] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 889.261301] env[61857]: DEBUG oslo_vmware.rw_handles [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 889.261301] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52efa3b0-0592-e821-7b70-eb4577627aba" [ 889.261301] env[61857]: _type = "HttpNfcLease" [ 889.261301] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 889.262057] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0441ef53-d166-412a-bd3b-cc8915e0621f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.273068] env[61857]: DEBUG oslo_vmware.rw_handles [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d00b8-db09-2fd5-dc32-2a1451adfd74/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 889.273068] env[61857]: DEBUG oslo_vmware.rw_handles [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d00b8-db09-2fd5-dc32-2a1451adfd74/disk-0.vmdk. {{(pid=61857) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 889.339634] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.045s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.343362] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cc8ffff5-5b32-4a40-970a-75ef472c86c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.345109] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.184s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.346862] env[61857]: INFO nova.compute.claims [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.369773] env[61857]: INFO nova.scheduler.client.report [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted allocations for instance 3fb5d24b-1767-43d9-bed5-833121962adb [ 889.534904] env[61857]: DEBUG oslo_vmware.api [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951172, 'name': PowerOffVM_Task, 'duration_secs': 0.189566} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.535197] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 889.535376] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 889.535947] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee169449-58e7-4c77-94d7-3d364cd811cc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.545894] env[61857]: DEBUG nova.network.neutron [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 889.589317] env[61857]: DEBUG oslo_vmware.api [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951173, 'name': PowerOffVM_Task, 'duration_secs': 0.250427} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.589607] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 889.589783] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 889.590049] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f320d21d-b9ec-4bea-9c1b-5df508e4bb5e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.673407] env[61857]: DEBUG nova.network.neutron [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updated VIF entry in instance network info cache for port 25a85668-66ff-4cf6-8cef-68b4d1a2e828. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 889.673871] env[61857]: DEBUG nova.network.neutron [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updating instance_info_cache with network_info: [{"id": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "address": "fa:16:3e:40:07:f3", "network": {"id": "c8dc3d6e-ad27-40ee-a7d9-3884fe81c2b8", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1462877859-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b51cf69f1c947c89a3969daba09361b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afd3feb3-ffcc-4499-a2c2-eb6a48aefde9", "external-id": "nsx-vlan-transportzone-22", "segmentation_id": 22, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25a85668-66", "ovs_interfaceid": "25a85668-66ff-4cf6-8cef-68b4d1a2e828", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.724205] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 889.724205] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 889.724205] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Deleting the datastore file [datastore2] 7a9252da-2584-40ed-9d28-ca7341ed5165 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.724588] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6ed2868-9137-4afa-b4f9-e89c85471555 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.731231] env[61857]: DEBUG nova.network.neutron [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Updating instance_info_cache with network_info: [{"id": "4ba4d151-0e84-4f6d-aee1-6843bae4fe64", "address": "fa:16:3e:fd:42:05", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ba4d151-0e", "ovs_interfaceid": "4ba4d151-0e84-4f6d-aee1-6843bae4fe64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.735038] env[61857]: DEBUG oslo_vmware.api [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 889.735038] env[61857]: value = "task-951176" [ 889.735038] env[61857]: _type = "Task" [ 889.735038] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.743591] env[61857]: DEBUG oslo_vmware.api [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951176, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.762139] env[61857]: DEBUG nova.compute.manager [req-efaf042a-d09b-48e7-aadc-de0d40065ef7 req-4c5d12f6-513d-48dc-a456-a87519d9779c service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Received event network-changed-4ba4d151-0e84-4f6d-aee1-6843bae4fe64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 889.762402] env[61857]: DEBUG nova.compute.manager [req-efaf042a-d09b-48e7-aadc-de0d40065ef7 req-4c5d12f6-513d-48dc-a456-a87519d9779c service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Refreshing instance network info cache due to event network-changed-4ba4d151-0e84-4f6d-aee1-6843bae4fe64. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 889.763038] env[61857]: DEBUG oslo_concurrency.lockutils [req-efaf042a-d09b-48e7-aadc-de0d40065ef7 req-4c5d12f6-513d-48dc-a456-a87519d9779c service nova] Acquiring lock "refresh_cache-d979a8a1-a5b5-428e-9b60-7d827337218f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.882258] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e205a93d-388a-4f09-ad2f-29de69a7c82c tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "3fb5d24b-1767-43d9-bed5-833121962adb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.089s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.181354] env[61857]: DEBUG oslo_concurrency.lockutils [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] Releasing lock "refresh_cache-24d3d515-e03a-4b4e-bb8e-bc18537125ac" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.181518] env[61857]: DEBUG nova.compute.manager [req-62fade9f-e99d-4a82-bc2a-35964d831489 req-202798f0-c93d-44df-9726-a5f26f592949 service nova] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Received event network-vif-deleted-7ede4f79-7428-48b0-a811-e8c064d224f0 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.242347] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "refresh_cache-d979a8a1-a5b5-428e-9b60-7d827337218f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.242740] env[61857]: DEBUG nova.compute.manager [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Instance network_info: |[{"id": "4ba4d151-0e84-4f6d-aee1-6843bae4fe64", "address": "fa:16:3e:fd:42:05", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ba4d151-0e", "ovs_interfaceid": "4ba4d151-0e84-4f6d-aee1-6843bae4fe64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 890.242955] env[61857]: DEBUG oslo_concurrency.lockutils [req-efaf042a-d09b-48e7-aadc-de0d40065ef7 req-4c5d12f6-513d-48dc-a456-a87519d9779c service nova] Acquired lock "refresh_cache-d979a8a1-a5b5-428e-9b60-7d827337218f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.243329] env[61857]: DEBUG nova.network.neutron [req-efaf042a-d09b-48e7-aadc-de0d40065ef7 req-4c5d12f6-513d-48dc-a456-a87519d9779c service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Refreshing network info cache for port 4ba4d151-0e84-4f6d-aee1-6843bae4fe64 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 890.246128] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:42:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ba4d151-0e84-4f6d-aee1-6843bae4fe64', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.260136] env[61857]: DEBUG oslo.service.loopingcall [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.262101] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 890.262475] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-923d499f-12e0-4096-a552-a84a316793af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.299656] env[61857]: DEBUG oslo_vmware.api [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951176, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17089} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.300694] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.301056] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 890.301397] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 890.301725] env[61857]: INFO nova.compute.manager [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Took 1.24 seconds to destroy the instance on the hypervisor. [ 890.302167] env[61857]: DEBUG oslo.service.loopingcall [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.302836] env[61857]: DEBUG nova.compute.manager [-] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.303013] env[61857]: DEBUG nova.network.neutron [-] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 890.311178] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.311178] env[61857]: value = "task-951177" [ 890.311178] env[61857]: _type = "Task" [ 890.311178] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.329110] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951177, 'name': CreateVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.352798] env[61857]: DEBUG oslo_vmware.rw_handles [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Completed reading data from the image iterator. {{(pid=61857) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 890.353128] env[61857]: DEBUG oslo_vmware.rw_handles [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d00b8-db09-2fd5-dc32-2a1451adfd74/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 890.354064] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d395243-26b4-43bd-bdf6-ef52eeece3b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.366648] env[61857]: DEBUG oslo_vmware.rw_handles [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d00b8-db09-2fd5-dc32-2a1451adfd74/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 890.367015] env[61857]: DEBUG oslo_vmware.rw_handles [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d00b8-db09-2fd5-dc32-2a1451adfd74/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 890.367212] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ff073da3-71fd-48f9-82b7-ab3d22c8f6d4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.607632] env[61857]: DEBUG oslo_vmware.rw_handles [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d00b8-db09-2fd5-dc32-2a1451adfd74/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 890.607911] env[61857]: INFO nova.virt.vmwareapi.images [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Downloaded image file data de6910ec-d5ea-4309-9229-4103bb53c96c [ 890.609090] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c5bcae-d856-467b-a8ab-138e407d293d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.630936] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e9f1e0d-6621-4783-9e8f-199160f4e462 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.666882] env[61857]: INFO nova.virt.vmwareapi.images [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] The imported VM was unregistered [ 890.669516] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Caching image {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 890.669744] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Creating directory with path [datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.670052] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1433c64d-a812-4698-a78f-2c82aecf741b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.684390] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Created directory with path [datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.684642] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f/OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f.vmdk to [datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c/de6910ec-d5ea-4309-9229-4103bb53c96c.vmdk. {{(pid=61857) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 890.685022] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-2701e645-f8b9-4da2-8627-1da2026ba990 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.693621] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 890.693621] env[61857]: value = "task-951179" [ 890.693621] env[61857]: _type = "Task" [ 890.693621] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.703692] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "16b447b5-426a-4478-9d44-ae32b41dee50" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.703975] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.709465] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951179, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.763112] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bad493-a97c-449b-a5d6-9b4a5566c69b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.771220] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a713dd-534f-4ab7-b44e-d12cd68602c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.805112] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab22a56-089e-4e30-95a4-457835de090d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.812398] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c1adfc-f43b-4642-b1bf-594f8bca66fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.825095] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951177, 'name': CreateVM_Task, 'duration_secs': 0.354146} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.834366] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 890.835185] env[61857]: DEBUG nova.compute.provider_tree [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.837018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.837275] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.837652] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.838157] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7c07bb7-33e8-463f-add1-17d900769a4d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.843840] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 890.843840] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52daea7d-6a91-45ec-9354-ea01b8901ded" [ 890.843840] env[61857]: _type = "Task" [ 890.843840] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.853095] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52daea7d-6a91-45ec-9354-ea01b8901ded, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.013491] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 891.013732] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 891.013913] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Deleting the datastore file [datastore1] ddd4c436-405f-49f7-8c9b-de3b71725f63 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.014221] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8838095-b4b7-4003-97fd-8f8efeab3272 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.026464] env[61857]: DEBUG oslo_vmware.api [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for the task: (returnval){ [ 891.026464] env[61857]: value = "task-951180" [ 891.026464] env[61857]: _type = "Task" [ 891.026464] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.039631] env[61857]: DEBUG oslo_vmware.api [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951180, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.104715] env[61857]: DEBUG nova.network.neutron [req-efaf042a-d09b-48e7-aadc-de0d40065ef7 req-4c5d12f6-513d-48dc-a456-a87519d9779c service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Updated VIF entry in instance network info cache for port 4ba4d151-0e84-4f6d-aee1-6843bae4fe64. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 891.105335] env[61857]: DEBUG nova.network.neutron [req-efaf042a-d09b-48e7-aadc-de0d40065ef7 req-4c5d12f6-513d-48dc-a456-a87519d9779c service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Updating instance_info_cache with network_info: [{"id": "4ba4d151-0e84-4f6d-aee1-6843bae4fe64", "address": "fa:16:3e:fd:42:05", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ba4d151-0e", "ovs_interfaceid": "4ba4d151-0e84-4f6d-aee1-6843bae4fe64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.180456] env[61857]: DEBUG nova.network.neutron [-] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.209653] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951179, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.210893] env[61857]: DEBUG nova.compute.utils [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.340017] env[61857]: DEBUG nova.scheduler.client.report [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.359483] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52daea7d-6a91-45ec-9354-ea01b8901ded, 'name': SearchDatastore_Task, 'duration_secs': 0.010263} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.359860] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.360158] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.360481] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.360642] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.360834] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.361389] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ca5afa4-41f0-497f-87e7-99f9fa7f6110 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.376298] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.376559] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 891.377662] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b4565fa-f46f-49f4-b906-348fbb187f76 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.385150] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 891.385150] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]528def06-6ae4-8e52-e31b-17175810b785" [ 891.385150] env[61857]: _type = "Task" [ 891.385150] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.397387] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528def06-6ae4-8e52-e31b-17175810b785, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.539294] env[61857]: DEBUG oslo_vmware.api [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951180, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.609606] env[61857]: DEBUG oslo_concurrency.lockutils [req-efaf042a-d09b-48e7-aadc-de0d40065ef7 req-4c5d12f6-513d-48dc-a456-a87519d9779c service nova] Releasing lock "refresh_cache-d979a8a1-a5b5-428e-9b60-7d827337218f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.683669] env[61857]: INFO nova.compute.manager [-] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Took 1.38 seconds to deallocate network for instance. [ 891.708536] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951179, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.716515] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.832693] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "992bcab3-7396-495f-96a1-e99ecd13c961" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.833074] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "992bcab3-7396-495f-96a1-e99ecd13c961" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.845371] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.845985] env[61857]: DEBUG nova.compute.manager [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.848984] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.104s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.849310] env[61857]: DEBUG nova.objects.instance [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lazy-loading 'resources' on Instance uuid ddeed65b-9003-443f-8b2b-0756fbe7d234 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.902765] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528def06-6ae4-8e52-e31b-17175810b785, 'name': SearchDatastore_Task, 'duration_secs': 0.022953} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.903866] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44147867-0daa-4088-994f-b0d8fd1183ef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.912778] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 891.912778] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b812ff-fd29-abd9-3d9c-593a1485fe17" [ 891.912778] env[61857]: _type = "Task" [ 891.912778] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.924124] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b812ff-fd29-abd9-3d9c-593a1485fe17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.044919] env[61857]: DEBUG oslo_vmware.api [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951180, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.165123] env[61857]: DEBUG nova.compute.manager [req-0e38631c-102e-44e0-99ab-82cd39f282be req-c413cd2a-9e18-4248-9901-9673b284111b service nova] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Received event network-vif-deleted-3a8e7408-98e0-411a-8413-0968ac906574 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.190685] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.207858] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951179, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.335822] env[61857]: DEBUG nova.compute.manager [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 892.357907] env[61857]: DEBUG nova.compute.utils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.359373] env[61857]: DEBUG nova.compute.manager [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.359690] env[61857]: DEBUG nova.network.neutron [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 892.428785] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b812ff-fd29-abd9-3d9c-593a1485fe17, 'name': SearchDatastore_Task, 'duration_secs': 0.021326} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.429761] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.429761] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] d979a8a1-a5b5-428e-9b60-7d827337218f/d979a8a1-a5b5-428e-9b60-7d827337218f.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 892.429761] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9bbd464e-53d3-431e-9ab8-71c4b0508ed5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.442997] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 892.442997] env[61857]: value = "task-951181" [ 892.442997] env[61857]: _type = "Task" [ 892.442997] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.451232] env[61857]: DEBUG nova.policy [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c21b2b2013cd40a69cc7a958a821e377', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ff67efd6404381a52a4001ec774a72', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.461062] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951181, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.542557] env[61857]: DEBUG oslo_vmware.api [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951180, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.712076] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951179, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.744039] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfe96bb-2f0f-4f2b-926f-65e96be5c094 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.754397] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d52aa9f-c414-4eaa-8111-b466d2ccedd8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.798070] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1792d7d8-5475-458f-b766-b1e0a3b1acc0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.803173] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "16b447b5-426a-4478-9d44-ae32b41dee50" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.803484] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.803735] env[61857]: INFO nova.compute.manager [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Attaching volume b8ef1d4a-6429-449a-942d-0ea7be737467 to /dev/sdc [ 892.816425] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6be8131-fcea-4f8e-89cc-8bddfda18497 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.833912] env[61857]: DEBUG nova.compute.provider_tree [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.836055] env[61857]: DEBUG nova.network.neutron [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Successfully created port: 639da037-e9cb-4624-994d-e7faab16e538 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.849666] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91687d5b-c475-44ed-ab05-857ab81e6e10 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.858253] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.861271] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c292458e-9952-40b6-9f5d-554d2a799e34 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.864896] env[61857]: DEBUG nova.compute.manager [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.886063] env[61857]: DEBUG nova.virt.block_device [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Updating existing volume attachment record: d2fe2a28-d956-468a-b1d6-521a8701a759 {{(pid=61857) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 892.956549] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951181, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.042403] env[61857]: DEBUG oslo_vmware.api [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Task: {'id': task-951180, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.662086} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.042805] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.042870] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 893.043028] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 893.043228] env[61857]: INFO nova.compute.manager [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Took 4.04 seconds to destroy the instance on the hypervisor. [ 893.043493] env[61857]: DEBUG oslo.service.loopingcall [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.043705] env[61857]: DEBUG nova.compute.manager [-] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 893.043806] env[61857]: DEBUG nova.network.neutron [-] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 893.206624] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951179, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.457981} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.208784] env[61857]: INFO nova.virt.vmwareapi.ds_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f/OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f.vmdk to [datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c/de6910ec-d5ea-4309-9229-4103bb53c96c.vmdk. [ 893.208999] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Cleaning up location [datastore1] OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 893.209184] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_174adeec-f36d-4bad-a314-beaedf0c288f {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.209478] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66086ec7-6742-42ac-b15a-898f4ca07274 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.216673] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 893.216673] env[61857]: value = "task-951183" [ 893.216673] env[61857]: _type = "Task" [ 893.216673] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.224879] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951183, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.339951] env[61857]: DEBUG nova.scheduler.client.report [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.454273] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951181, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.680446} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.454627] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] d979a8a1-a5b5-428e-9b60-7d827337218f/d979a8a1-a5b5-428e-9b60-7d827337218f.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 893.454841] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.455118] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bebc7ffe-656a-45cc-9ae3-c86d6cb71f35 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.466132] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 893.466132] env[61857]: value = "task-951184" [ 893.466132] env[61857]: _type = "Task" [ 893.466132] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.475554] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951184, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.727927] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951183, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.041474} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.728122] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.728328] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "[datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c/de6910ec-d5ea-4309-9229-4103bb53c96c.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.728582] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c/de6910ec-d5ea-4309-9229-4103bb53c96c.vmdk to [datastore1] 33cb5bbc-926d-42ee-b483-8d1e24707e40/33cb5bbc-926d-42ee-b483-8d1e24707e40.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 893.728838] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8625f79c-6192-4ecd-845d-e21737bd072e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.736273] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 893.736273] env[61857]: value = "task-951185" [ 893.736273] env[61857]: _type = "Task" [ 893.736273] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.744418] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951185, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.846373] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.997s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.848306] env[61857]: DEBUG oslo_concurrency.lockutils [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.947s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.848557] env[61857]: DEBUG nova.objects.instance [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lazy-loading 'resources' on Instance uuid 6c4c1276-ae80-4f37-9ef9-4872dd415d24 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.868700] env[61857]: INFO nova.scheduler.client.report [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleted allocations for instance ddeed65b-9003-443f-8b2b-0756fbe7d234 [ 893.874258] env[61857]: DEBUG nova.compute.manager [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.901404] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.901727] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.901937] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.902184] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.902656] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.902880] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.903208] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.903333] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.903544] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.903687] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.903869] env[61857]: DEBUG nova.virt.hardware [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.904764] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d588513-da9a-4323-8a12-1abb978b7575 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.913270] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3913136-a8eb-4086-91b8-9755600d06bb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.935124] env[61857]: DEBUG nova.network.neutron [-] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.976373] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951184, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073386} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.976748] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.977749] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac87dbd-74b3-4000-8dd8-0b56adf713dc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.001610] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] d979a8a1-a5b5-428e-9b60-7d827337218f/d979a8a1-a5b5-428e-9b60-7d827337218f.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.001937] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-678a1b2c-c573-4822-9e23-ba8ee96be57c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.022580] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 894.022580] env[61857]: value = "task-951186" [ 894.022580] env[61857]: _type = "Task" [ 894.022580] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.030629] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951186, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.202677] env[61857]: DEBUG nova.compute.manager [req-b857ef3f-8b88-4bfd-a6d2-4892368d83f5 req-42ccb716-f8b8-4b86-818c-75884f439783 service nova] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Received event network-vif-deleted-ec7e2b6d-16bc-4dc7-9355-f76d47ed088a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 894.252640] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951185, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.381133] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f88599d3-6875-49b5-af95-d3557657bb7d tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "ddeed65b-9003-443f-8b2b-0756fbe7d234" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.820s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.438930] env[61857]: INFO nova.compute.manager [-] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Took 1.39 seconds to deallocate network for instance. [ 894.535178] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951186, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.710071] env[61857]: DEBUG nova.network.neutron [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Successfully updated port: 639da037-e9cb-4624-994d-e7faab16e538 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.755174] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951185, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.789321] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593b66eb-9fde-4cea-9b27-f06c981224e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.800301] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fdeabc-745e-44cb-b699-2645932f6f8c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.840677] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85bae616-ae7b-4fe9-a4ba-1a518557e559 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.852736] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb67197-d845-4d3c-9357-53ed42a88a24 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.870863] env[61857]: DEBUG nova.compute.provider_tree [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.950034] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.038810] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951186, 'name': ReconfigVM_Task, 'duration_secs': 0.744301} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.039136] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Reconfigured VM instance instance-00000050 to attach disk [datastore2] d979a8a1-a5b5-428e-9b60-7d827337218f/d979a8a1-a5b5-428e-9b60-7d827337218f.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.039813] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79a33609-2bbc-4ce1-a1d1-f6b01d051a17 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.050903] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 895.050903] env[61857]: value = "task-951187" [ 895.050903] env[61857]: _type = "Task" [ 895.050903] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.061482] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951187, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.217812] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "refresh_cache-0de8e67e-24b9-48d0-ac90-b5ec5b93e243" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.218077] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquired lock "refresh_cache-0de8e67e-24b9-48d0-ac90-b5ec5b93e243" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.218362] env[61857]: DEBUG nova.network.neutron [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 895.252120] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951185, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.374153] env[61857]: DEBUG nova.scheduler.client.report [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.563750] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951187, 'name': Rename_Task, 'duration_secs': 0.271321} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.564199] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.564376] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f9e25ec-e7a5-4691-b876-be9d1963d80c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.575357] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 895.575357] env[61857]: value = "task-951189" [ 895.575357] env[61857]: _type = "Task" [ 895.575357] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.587016] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951189, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.752020] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951185, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.754358] env[61857]: DEBUG nova.network.neutron [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 895.879475] env[61857]: DEBUG oslo_concurrency.lockutils [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.031s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.881850] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.588s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.882361] env[61857]: DEBUG nova.objects.instance [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lazy-loading 'resources' on Instance uuid c27f6f5b-6423-4eee-be53-dc48c93df893 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.908745] env[61857]: INFO nova.scheduler.client.report [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Deleted allocations for instance 6c4c1276-ae80-4f37-9ef9-4872dd415d24 [ 895.913120] env[61857]: DEBUG nova.network.neutron [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Updating instance_info_cache with network_info: [{"id": "639da037-e9cb-4624-994d-e7faab16e538", "address": "fa:16:3e:4a:86:b0", "network": {"id": "aa2712a7-cfd9-4139-8b6d-508cc240cee4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-131420680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ff67efd6404381a52a4001ec774a72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfbfc55d-8126-40dd-998e-8600ea92f97c", "external-id": "nsx-vlan-transportzone-650", "segmentation_id": 650, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap639da037-e9", "ovs_interfaceid": "639da037-e9cb-4624-994d-e7faab16e538", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.090435] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951189, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.232314] env[61857]: DEBUG nova.compute.manager [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Received event network-vif-plugged-639da037-e9cb-4624-994d-e7faab16e538 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.232314] env[61857]: DEBUG oslo_concurrency.lockutils [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] Acquiring lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.232547] env[61857]: DEBUG oslo_concurrency.lockutils [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.232601] env[61857]: DEBUG oslo_concurrency.lockutils [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.232769] env[61857]: DEBUG nova.compute.manager [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] No waiting events found dispatching network-vif-plugged-639da037-e9cb-4624-994d-e7faab16e538 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 896.233034] env[61857]: WARNING nova.compute.manager [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Received unexpected event network-vif-plugged-639da037-e9cb-4624-994d-e7faab16e538 for instance with vm_state building and task_state spawning. [ 896.233117] env[61857]: DEBUG nova.compute.manager [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Received event network-changed-639da037-e9cb-4624-994d-e7faab16e538 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.233293] env[61857]: DEBUG nova.compute.manager [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Refreshing instance network info cache due to event network-changed-639da037-e9cb-4624-994d-e7faab16e538. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 896.233293] env[61857]: DEBUG oslo_concurrency.lockutils [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] Acquiring lock "refresh_cache-0de8e67e-24b9-48d0-ac90-b5ec5b93e243" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.251342] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951185, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.417940] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Releasing lock "refresh_cache-0de8e67e-24b9-48d0-ac90-b5ec5b93e243" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.418365] env[61857]: DEBUG nova.compute.manager [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Instance network_info: |[{"id": "639da037-e9cb-4624-994d-e7faab16e538", "address": "fa:16:3e:4a:86:b0", "network": {"id": "aa2712a7-cfd9-4139-8b6d-508cc240cee4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-131420680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ff67efd6404381a52a4001ec774a72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfbfc55d-8126-40dd-998e-8600ea92f97c", "external-id": "nsx-vlan-transportzone-650", "segmentation_id": 650, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap639da037-e9", "ovs_interfaceid": "639da037-e9cb-4624-994d-e7faab16e538", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 896.418665] env[61857]: DEBUG oslo_concurrency.lockutils [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] Acquired lock "refresh_cache-0de8e67e-24b9-48d0-ac90-b5ec5b93e243" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.418854] env[61857]: DEBUG nova.network.neutron [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Refreshing network info cache for port 639da037-e9cb-4624-994d-e7faab16e538 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 896.420479] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:86:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bfbfc55d-8126-40dd-998e-8600ea92f97c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '639da037-e9cb-4624-994d-e7faab16e538', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.430531] env[61857]: DEBUG oslo.service.loopingcall [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.433952] env[61857]: DEBUG oslo_concurrency.lockutils [None req-44d33593-53b1-46f3-82ea-a6e891af8523 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "6c4c1276-ae80-4f37-9ef9-4872dd415d24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.652s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.437963] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 896.438186] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e03c9374-87a1-453b-82e2-0ca3d81fa86e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.465500] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.465500] env[61857]: value = "task-951190" [ 896.465500] env[61857]: _type = "Task" [ 896.465500] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.477175] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951190, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.565928] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.566239] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.566346] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.566517] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.566689] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.569011] env[61857]: INFO nova.compute.manager [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Terminating instance [ 896.573635] env[61857]: DEBUG nova.compute.manager [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 896.573842] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 896.574690] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9d0914-1950-4fdd-9cc4-02f1a5efe783 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.589450] env[61857]: DEBUG oslo_vmware.api [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951189, 'name': PowerOnVM_Task, 'duration_secs': 0.712353} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.591724] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.591944] env[61857]: INFO nova.compute.manager [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Took 9.42 seconds to spawn the instance on the hypervisor. [ 896.592153] env[61857]: DEBUG nova.compute.manager [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.595594] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 896.595594] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b46cfaf-8530-460a-89e0-2d9d5dc8ba31 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.595594] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f16b178-4ffb-4b14-9770-d5daa31bd44a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.610118] env[61857]: DEBUG oslo_vmware.api [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 896.610118] env[61857]: value = "task-951191" [ 896.610118] env[61857]: _type = "Task" [ 896.610118] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.621709] env[61857]: DEBUG oslo_vmware.api [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951191, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.739146] env[61857]: DEBUG nova.network.neutron [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Updated VIF entry in instance network info cache for port 639da037-e9cb-4624-994d-e7faab16e538. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 896.739521] env[61857]: DEBUG nova.network.neutron [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Updating instance_info_cache with network_info: [{"id": "639da037-e9cb-4624-994d-e7faab16e538", "address": "fa:16:3e:4a:86:b0", "network": {"id": "aa2712a7-cfd9-4139-8b6d-508cc240cee4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-131420680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ff67efd6404381a52a4001ec774a72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfbfc55d-8126-40dd-998e-8600ea92f97c", "external-id": "nsx-vlan-transportzone-650", "segmentation_id": 650, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap639da037-e9", "ovs_interfaceid": "639da037-e9cb-4624-994d-e7faab16e538", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.750968] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951185, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.710469} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.752827] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/de6910ec-d5ea-4309-9229-4103bb53c96c/de6910ec-d5ea-4309-9229-4103bb53c96c.vmdk to [datastore1] 33cb5bbc-926d-42ee-b483-8d1e24707e40/33cb5bbc-926d-42ee-b483-8d1e24707e40.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 896.754020] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4676ef8-7518-4bc7-918c-4640bffe3b20 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.757235] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35c0ed6-2d9e-42fc-9268-c97d009b2f95 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.774927] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eda00c2-2250-4d13-8858-4d20c09a696a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.787043] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 33cb5bbc-926d-42ee-b483-8d1e24707e40/33cb5bbc-926d-42ee-b483-8d1e24707e40.vmdk or device None with type streamOptimized {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.787363] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4aadebb6-d1d7-445f-9dda-1e7bcab141ab {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.833150] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba05637-f972-4283-b264-aa3faa997da6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.836057] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 896.836057] env[61857]: value = "task-951192" [ 896.836057] env[61857]: _type = "Task" [ 896.836057] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.843751] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684fa5f5-530e-4f34-be55-2b8d379c46c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.851354] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951192, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.863289] env[61857]: DEBUG nova.compute.provider_tree [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.977213] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951190, 'name': CreateVM_Task, 'duration_secs': 0.457299} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.977373] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 896.978143] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.978472] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.978860] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.979141] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81048de4-0bdd-4cc3-a78c-351462fd985a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.983995] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 896.983995] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52139551-a617-dfc6-b7ee-b6b7c1b150d7" [ 896.983995] env[61857]: _type = "Task" [ 896.983995] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.991730] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52139551-a617-dfc6-b7ee-b6b7c1b150d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.122200] env[61857]: INFO nova.compute.manager [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Took 40.17 seconds to build instance. [ 897.126662] env[61857]: DEBUG oslo_vmware.api [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951191, 'name': PowerOffVM_Task, 'duration_secs': 0.276631} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.127204] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 897.127465] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 897.127748] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a5de3d31-0a9e-4012-b3d5-ade152f56d71 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.202144] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 897.202423] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 897.202641] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Deleting the datastore file [datastore1] 42d38f5e-628a-4030-85e1-3ec0595cf3c8 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.202921] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-168ead04-b41a-4048-9f26-06e374363011 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.210742] env[61857]: DEBUG oslo_vmware.api [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for the task: (returnval){ [ 897.210742] env[61857]: value = "task-951194" [ 897.210742] env[61857]: _type = "Task" [ 897.210742] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.219075] env[61857]: DEBUG oslo_vmware.api [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.245883] env[61857]: DEBUG oslo_concurrency.lockutils [req-120fe5d7-2b1f-4574-8aef-15ff6ee841ef req-b71db9eb-d2b6-4f18-beb4-a942add06d05 service nova] Releasing lock "refresh_cache-0de8e67e-24b9-48d0-ac90-b5ec5b93e243" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.347133] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951192, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.366303] env[61857]: DEBUG nova.scheduler.client.report [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.450293] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Volume attach. Driver type: vmdk {{(pid=61857) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 897.450560] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214153', 'volume_id': 'b8ef1d4a-6429-449a-942d-0ea7be737467', 'name': 'volume-b8ef1d4a-6429-449a-942d-0ea7be737467', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '16b447b5-426a-4478-9d44-ae32b41dee50', 'attached_at': '', 'detached_at': '', 'volume_id': 'b8ef1d4a-6429-449a-942d-0ea7be737467', 'serial': 'b8ef1d4a-6429-449a-942d-0ea7be737467'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 897.451465] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8c7d69-a3d5-4e4e-85ee-a5ecb3df2e8c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.470062] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d354efe-440a-4b23-bcb8-f315cbdbdebd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.497494] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] volume-b8ef1d4a-6429-449a-942d-0ea7be737467/volume-b8ef1d4a-6429-449a-942d-0ea7be737467.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.500805] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22031f8e-5b29-4962-bb77-7ed5b5e4cbe5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.520322] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52139551-a617-dfc6-b7ee-b6b7c1b150d7, 'name': SearchDatastore_Task, 'duration_secs': 0.020903} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.521912] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.522181] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 897.522459] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.522663] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.522902] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.523247] env[61857]: DEBUG oslo_vmware.api [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 897.523247] env[61857]: value = "task-951195" [ 897.523247] env[61857]: _type = "Task" [ 897.523247] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.523450] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9630da23-a54c-48fa-b9f8-4c9e2a739fbd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.533024] env[61857]: DEBUG oslo_vmware.api [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951195, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.542967] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.543201] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 897.543933] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f943b9d-8097-4253-800e-8bd0aa610542 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.549395] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 897.549395] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]529d8fd6-6ddd-03af-83a4-46f9ee532177" [ 897.549395] env[61857]: _type = "Task" [ 897.549395] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.557910] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529d8fd6-6ddd-03af-83a4-46f9ee532177, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.627901] env[61857]: DEBUG oslo_concurrency.lockutils [None req-04d5447b-2ad3-410e-9ad9-f7f631c0ae1d tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.709s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.721152] env[61857]: DEBUG oslo_vmware.api [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Task: {'id': task-951194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.455369} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.721407] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.721601] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 897.721809] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 897.721995] env[61857]: INFO nova.compute.manager [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 897.722271] env[61857]: DEBUG oslo.service.loopingcall [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.722478] env[61857]: DEBUG nova.compute.manager [-] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 897.722579] env[61857]: DEBUG nova.network.neutron [-] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 897.847470] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951192, 'name': ReconfigVM_Task, 'duration_secs': 0.911448} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.847774] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 33cb5bbc-926d-42ee-b483-8d1e24707e40/33cb5bbc-926d-42ee-b483-8d1e24707e40.vmdk or device None with type streamOptimized {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.848450] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39f776bc-0b1c-4701-948a-6d2f1626aafc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.856404] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 897.856404] env[61857]: value = "task-951196" [ 897.856404] env[61857]: _type = "Task" [ 897.856404] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.864787] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951196, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.872017] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.990s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.874284] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.287s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.874553] env[61857]: DEBUG nova.objects.instance [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'resources' on Instance uuid 96ff157c-4471-4a12-ad99-0aafd6c2dce5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.903752] env[61857]: INFO nova.scheduler.client.report [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Deleted allocations for instance c27f6f5b-6423-4eee-be53-dc48c93df893 [ 898.037168] env[61857]: DEBUG oslo_vmware.api [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951195, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.060276] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529d8fd6-6ddd-03af-83a4-46f9ee532177, 'name': SearchDatastore_Task, 'duration_secs': 0.01283} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.061115] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-411fece4-8764-4f24-bfb1-d9b3f2e9c595 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.069934] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 898.069934] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b885d0-4673-d06c-dabd-fb7119c7274d" [ 898.069934] env[61857]: _type = "Task" [ 898.069934] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.080819] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b885d0-4673-d06c-dabd-fb7119c7274d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.312316] env[61857]: DEBUG nova.compute.manager [req-53a29bfa-ed87-49a1-a2e6-34b89684138c req-7ca27fe4-5a6d-4c2a-b2ca-2344c4857981 service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Received event network-vif-deleted-467ad7db-bcf7-4b1c-8fa3-a3d03af32d88 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.312539] env[61857]: INFO nova.compute.manager [req-53a29bfa-ed87-49a1-a2e6-34b89684138c req-7ca27fe4-5a6d-4c2a-b2ca-2344c4857981 service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Neutron deleted interface 467ad7db-bcf7-4b1c-8fa3-a3d03af32d88; detaching it from the instance and deleting it from the info cache [ 898.312719] env[61857]: DEBUG nova.network.neutron [req-53a29bfa-ed87-49a1-a2e6-34b89684138c req-7ca27fe4-5a6d-4c2a-b2ca-2344c4857981 service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.371496] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951196, 'name': Rename_Task, 'duration_secs': 0.24052} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.371827] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 898.372391] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-415b909b-a1d0-42b9-a173-1aaf1aa1465a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.381830] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 898.381830] env[61857]: value = "task-951197" [ 898.381830] env[61857]: _type = "Task" [ 898.381830] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.390944] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951197, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.414321] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aed48a1d-cbcf-488d-af94-8800e276b9fd tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "c27f6f5b-6423-4eee-be53-dc48c93df893" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.142s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.537865] env[61857]: DEBUG oslo_vmware.api [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951195, 'name': ReconfigVM_Task, 'duration_secs': 0.59298} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.541388] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Reconfigured VM instance instance-00000042 to attach disk [datastore2] volume-b8ef1d4a-6429-449a-942d-0ea7be737467/volume-b8ef1d4a-6429-449a-942d-0ea7be737467.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.545873] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d34f64c-a38e-4e5c-94f0-bfc41d55ed4a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.563767] env[61857]: DEBUG oslo_vmware.api [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 898.563767] env[61857]: value = "task-951198" [ 898.563767] env[61857]: _type = "Task" [ 898.563767] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.567447] env[61857]: DEBUG nova.network.neutron [-] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.581441] env[61857]: DEBUG oslo_vmware.api [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951198, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.589179] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b885d0-4673-d06c-dabd-fb7119c7274d, 'name': SearchDatastore_Task, 'duration_secs': 0.016145} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.589179] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.589179] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 0de8e67e-24b9-48d0-ac90-b5ec5b93e243/0de8e67e-24b9-48d0-ac90-b5ec5b93e243.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 898.589179] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5ba6e16-edd4-408c-b6cd-f62a88bdf14a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.596491] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 898.596491] env[61857]: value = "task-951199" [ 898.596491] env[61857]: _type = "Task" [ 898.596491] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.605161] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951199, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.711551] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162d01fe-1fdd-40d1-9bf3-6d64282dbe2b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.719943] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b59f74-e32e-467e-9aaa-505928d86f4b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.752201] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1400e92b-1afb-4a5d-951c-2b41ac30b5b9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.760938] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6871bf-9be6-4085-ae1a-73d7b2c13a91 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.776131] env[61857]: DEBUG nova.compute.provider_tree [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.815698] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d28b90a-4220-4349-8a20-5c76549ee2b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.826517] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29b7700-aeb2-41f9-bcac-f18328754388 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.861335] env[61857]: DEBUG nova.compute.manager [req-53a29bfa-ed87-49a1-a2e6-34b89684138c req-7ca27fe4-5a6d-4c2a-b2ca-2344c4857981 service nova] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Detach interface failed, port_id=467ad7db-bcf7-4b1c-8fa3-a3d03af32d88, reason: Instance 42d38f5e-628a-4030-85e1-3ec0595cf3c8 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 898.891841] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951197, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.071343] env[61857]: INFO nova.compute.manager [-] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Took 1.35 seconds to deallocate network for instance. [ 899.076598] env[61857]: DEBUG oslo_vmware.api [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.104052] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "d979a8a1-a5b5-428e-9b60-7d827337218f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.104320] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.109689] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951199, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505832} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.110299] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 0de8e67e-24b9-48d0-ac90-b5ec5b93e243/0de8e67e-24b9-48d0-ac90-b5ec5b93e243.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 899.110517] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 899.111114] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dbde2947-c27b-45d8-9b2a-231d36b200c5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.119950] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 899.119950] env[61857]: value = "task-951200" [ 899.119950] env[61857]: _type = "Task" [ 899.119950] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.131309] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951200, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.279016] env[61857]: DEBUG nova.scheduler.client.report [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.391937] env[61857]: DEBUG oslo_vmware.api [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951197, 'name': PowerOnVM_Task, 'duration_secs': 0.861985} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.392241] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 899.392451] env[61857]: INFO nova.compute.manager [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Took 17.04 seconds to spawn the instance on the hypervisor. [ 899.392639] env[61857]: DEBUG nova.compute.manager [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.393435] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d943855d-b8f3-4d4e-ada4-601126cd3a86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.577042] env[61857]: DEBUG oslo_vmware.api [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951198, 'name': ReconfigVM_Task, 'duration_secs': 0.606442} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.577042] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214153', 'volume_id': 'b8ef1d4a-6429-449a-942d-0ea7be737467', 'name': 'volume-b8ef1d4a-6429-449a-942d-0ea7be737467', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '16b447b5-426a-4478-9d44-ae32b41dee50', 'attached_at': '', 'detached_at': '', 'volume_id': 'b8ef1d4a-6429-449a-942d-0ea7be737467', 'serial': 'b8ef1d4a-6429-449a-942d-0ea7be737467'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 899.581776] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.612048] env[61857]: DEBUG nova.compute.utils [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.631133] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951200, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070433} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.631455] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.632377] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84948489-be07-40eb-a18c-5b57e951e1e3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.656835] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 0de8e67e-24b9-48d0-ac90-b5ec5b93e243/0de8e67e-24b9-48d0-ac90-b5ec5b93e243.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.656835] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c846be11-1b2b-4310-9d45-27af1aa04f94 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.677956] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 899.677956] env[61857]: value = "task-951201" [ 899.677956] env[61857]: _type = "Task" [ 899.677956] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.686961] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951201, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.784463] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.910s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.787058] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.416s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.788408] env[61857]: INFO nova.compute.claims [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.804236] env[61857]: INFO nova.scheduler.client.report [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleted allocations for instance 96ff157c-4471-4a12-ad99-0aafd6c2dce5 [ 899.912575] env[61857]: INFO nova.compute.manager [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Took 45.52 seconds to build instance. [ 900.115495] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.189048] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951201, 'name': ReconfigVM_Task, 'duration_secs': 0.262087} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.189376] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 0de8e67e-24b9-48d0-ac90-b5ec5b93e243/0de8e67e-24b9-48d0-ac90-b5ec5b93e243.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.190466] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27e4ac23-81c2-489d-aef9-c71827f7da6f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.197403] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 900.197403] env[61857]: value = "task-951202" [ 900.197403] env[61857]: _type = "Task" [ 900.197403] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.205700] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951202, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.312142] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bd2f3698-e181-4ad4-9527-c6e7ebdbdb10 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "96ff157c-4471-4a12-ad99-0aafd6c2dce5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.082s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.377862] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "33cb5bbc-926d-42ee-b483-8d1e24707e40" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.413745] env[61857]: DEBUG oslo_concurrency.lockutils [None req-38409aec-f084-4a68-9198-0a7eb57504a1 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "33cb5bbc-926d-42ee-b483-8d1e24707e40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.031s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.414084] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "33cb5bbc-926d-42ee-b483-8d1e24707e40" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.037s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.414495] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "33cb5bbc-926d-42ee-b483-8d1e24707e40-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.414536] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "33cb5bbc-926d-42ee-b483-8d1e24707e40-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.414722] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "33cb5bbc-926d-42ee-b483-8d1e24707e40-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.417244] env[61857]: INFO nova.compute.manager [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Terminating instance [ 900.419669] env[61857]: DEBUG nova.compute.manager [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 900.419901] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 900.421040] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3a2a35-ddff-4990-bc93-828928cfbd6e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.431462] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 900.431790] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfc1ec1a-01cb-4161-a0b0-833ea4199dea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.439366] env[61857]: DEBUG oslo_vmware.api [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 900.439366] env[61857]: value = "task-951203" [ 900.439366] env[61857]: _type = "Task" [ 900.439366] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.450584] env[61857]: DEBUG oslo_vmware.api [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951203, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.619182] env[61857]: DEBUG nova.objects.instance [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lazy-loading 'flavor' on Instance uuid 16b447b5-426a-4478-9d44-ae32b41dee50 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.710494] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951202, 'name': Rename_Task, 'duration_secs': 0.273023} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.714177] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 900.714177] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d4b14e9-6212-477f-a198-328eca1348f1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.721612] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 900.721612] env[61857]: value = "task-951204" [ 900.721612] env[61857]: _type = "Task" [ 900.721612] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.732431] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951204, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.950171] env[61857]: DEBUG oslo_vmware.api [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951203, 'name': PowerOffVM_Task, 'duration_secs': 0.289539} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.953848] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 900.954040] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 900.957925] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5fa835d6-a759-4881-9ee3-5118583fcc86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.045374] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 901.045634] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 901.045866] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleting the datastore file [datastore1] 33cb5bbc-926d-42ee-b483-8d1e24707e40 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.047382] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f5fc973-4f31-4093-b5c5-730c65d229a2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.056465] env[61857]: DEBUG oslo_vmware.api [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 901.056465] env[61857]: value = "task-951206" [ 901.056465] env[61857]: _type = "Task" [ 901.056465] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.065755] env[61857]: DEBUG oslo_vmware.api [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951206, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.128572] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb90c6e3-ec68-4354-9816-7c6ee0871753 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.325s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.150705] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a813cf92-1e30-4935-990e-d608c87b684d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.166320] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac5cb26-7be4-4130-a8b6-384b1e8dcf27 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.207228] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "d979a8a1-a5b5-428e-9b60-7d827337218f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.207532] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.207774] env[61857]: INFO nova.compute.manager [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Attaching volume 9647e67e-e4e4-4753-88a6-c144896c3f9c to /dev/sdb [ 901.211307] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6b2229-1cff-4aa9-88c7-bdfda6348b20 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.221690] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c85fe06-ca30-4245-839f-ce374c73f647 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.243481] env[61857]: DEBUG oslo_vmware.api [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951204, 'name': PowerOnVM_Task, 'duration_secs': 0.48778} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.244424] env[61857]: DEBUG nova.compute.provider_tree [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.247020] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 901.247020] env[61857]: INFO nova.compute.manager [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Took 7.37 seconds to spawn the instance on the hypervisor. [ 901.247020] env[61857]: DEBUG nova.compute.manager [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.247730] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c265e0d5-3672-43ae-b8eb-82463686e07c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.251735] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65d8f77-b0ae-447a-afe1-2a2987d4bb2f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.260058] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812c3794-5312-4887-ab54-d88facb10c65 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.279705] env[61857]: DEBUG nova.virt.block_device [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Updating existing volume attachment record: 52d273b5-4a96-4d10-9c3e-47933f2965df {{(pid=61857) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 901.461735] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "e88aed62-6561-4e29-b041-46b19cc5ef63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.462026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "e88aed62-6561-4e29-b041-46b19cc5ef63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.462828] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "e88aed62-6561-4e29-b041-46b19cc5ef63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.462828] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "e88aed62-6561-4e29-b041-46b19cc5ef63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.462828] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "e88aed62-6561-4e29-b041-46b19cc5ef63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.468205] env[61857]: INFO nova.compute.manager [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Terminating instance [ 901.468205] env[61857]: DEBUG nova.compute.manager [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.468205] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 901.469012] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d060bd65-13f1-4178-aa31-a5066cdda051 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.481562] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 901.481824] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be6a650e-d94c-4ad7-a22e-931820c0b018 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.489439] env[61857]: DEBUG oslo_vmware.api [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 901.489439] env[61857]: value = "task-951208" [ 901.489439] env[61857]: _type = "Task" [ 901.489439] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.498569] env[61857]: DEBUG oslo_vmware.api [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951208, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.567890] env[61857]: DEBUG oslo_vmware.api [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951206, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173213} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.567890] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.567890] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 901.567890] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 901.568134] env[61857]: INFO nova.compute.manager [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Took 1.15 seconds to destroy the instance on the hypervisor. [ 901.568218] env[61857]: DEBUG oslo.service.loopingcall [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.568421] env[61857]: DEBUG nova.compute.manager [-] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 901.568515] env[61857]: DEBUG nova.network.neutron [-] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 901.659764] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "16b447b5-426a-4478-9d44-ae32b41dee50" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.660942] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.755062] env[61857]: DEBUG nova.scheduler.client.report [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.776889] env[61857]: INFO nova.compute.manager [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Took 27.64 seconds to build instance. [ 902.006077] env[61857]: DEBUG oslo_vmware.api [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951208, 'name': PowerOffVM_Task, 'duration_secs': 0.217979} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.006077] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 902.006077] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 902.006077] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4195f59-7091-42fd-9c92-912f001c1970 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.139288] env[61857]: DEBUG nova.compute.manager [req-607666be-8ab7-4058-92a1-d2d9a58e6615 req-18b40d1a-8fdf-47c2-b0ee-4b3144d93aa2 service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Received event network-vif-deleted-11fe2d50-29b2-4bd7-90ba-b5d587f87ded {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.139519] env[61857]: INFO nova.compute.manager [req-607666be-8ab7-4058-92a1-d2d9a58e6615 req-18b40d1a-8fdf-47c2-b0ee-4b3144d93aa2 service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Neutron deleted interface 11fe2d50-29b2-4bd7-90ba-b5d587f87ded; detaching it from the instance and deleting it from the info cache [ 902.139709] env[61857]: DEBUG nova.network.neutron [req-607666be-8ab7-4058-92a1-d2d9a58e6615 req-18b40d1a-8fdf-47c2-b0ee-4b3144d93aa2 service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.165227] env[61857]: INFO nova.compute.manager [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Detaching volume 53af836e-3d73-4b7e-86c0-3d1641d68f9b [ 902.169118] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 902.169374] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 902.169706] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Deleting the datastore file [datastore2] e88aed62-6561-4e29-b041-46b19cc5ef63 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.169809] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18872ee0-e829-4d84-a184-7325ea35506b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.180699] env[61857]: DEBUG oslo_vmware.api [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for the task: (returnval){ [ 902.180699] env[61857]: value = "task-951212" [ 902.180699] env[61857]: _type = "Task" [ 902.180699] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.190496] env[61857]: DEBUG oslo_vmware.api [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951212, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.207933] env[61857]: INFO nova.virt.block_device [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Attempting to driver detach volume 53af836e-3d73-4b7e-86c0-3d1641d68f9b from mountpoint /dev/sdb [ 902.208353] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Volume detach. Driver type: vmdk {{(pid=61857) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 902.208614] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214149', 'volume_id': '53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'name': 'volume-53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '16b447b5-426a-4478-9d44-ae32b41dee50', 'attached_at': '', 'detached_at': '', 'volume_id': '53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'serial': '53af836e-3d73-4b7e-86c0-3d1641d68f9b'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 902.211031] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7c6776-9767-4308-bfd4-be2a781dac8d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.240732] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8524f5f2-c361-4bef-82f7-1e41c15bbe04 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.248640] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e71e75-b68b-4ac5-86ad-ecf1abe0a581 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.277550] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.278209] env[61857]: DEBUG nova.compute.manager [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 902.282151] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9efa3c25-c825-44af-951f-c90f67a44898 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.158s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.282151] env[61857]: DEBUG oslo_concurrency.lockutils [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.513s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.282151] env[61857]: DEBUG nova.objects.instance [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lazy-loading 'resources' on Instance uuid 6957dc98-8c60-4fdd-83dd-be2f13825c6d {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.284368] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb22369-c9de-4db6-8be8-e8842f6879ab {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.310256] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] The volume has not been displaced from its original location: [datastore1] volume-53af836e-3d73-4b7e-86c0-3d1641d68f9b/volume-53af836e-3d73-4b7e-86c0-3d1641d68f9b.vmdk. No consolidation needed. {{(pid=61857) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 902.318825] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Reconfiguring VM instance instance-00000042 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 902.319757] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a66a78c-457f-4539-b045-b7bef7c6521b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.348693] env[61857]: DEBUG oslo_vmware.api [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 902.348693] env[61857]: value = "task-951213" [ 902.348693] env[61857]: _type = "Task" [ 902.348693] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.359614] env[61857]: DEBUG oslo_vmware.api [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951213, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.587547] env[61857]: DEBUG nova.network.neutron [-] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.642655] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9fdaa9d5-6984-4f63-9118-ccb970dfcf05 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.664889] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea196e15-6790-41f6-8724-358f46a9b6cd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.691688] env[61857]: DEBUG oslo_vmware.api [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Task: {'id': task-951212, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198569} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.703360] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.703360] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 902.703360] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 902.703360] env[61857]: INFO nova.compute.manager [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Took 1.23 seconds to destroy the instance on the hypervisor. [ 902.703360] env[61857]: DEBUG oslo.service.loopingcall [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.703360] env[61857]: DEBUG nova.compute.manager [req-607666be-8ab7-4058-92a1-d2d9a58e6615 req-18b40d1a-8fdf-47c2-b0ee-4b3144d93aa2 service nova] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Detach interface failed, port_id=11fe2d50-29b2-4bd7-90ba-b5d587f87ded, reason: Instance 33cb5bbc-926d-42ee-b483-8d1e24707e40 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 902.703360] env[61857]: DEBUG nova.compute.manager [-] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 902.703360] env[61857]: DEBUG nova.network.neutron [-] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 902.794739] env[61857]: DEBUG nova.compute.utils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.796565] env[61857]: DEBUG nova.compute.manager [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 902.796892] env[61857]: DEBUG nova.network.neutron [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 902.862495] env[61857]: DEBUG oslo_vmware.api [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951213, 'name': ReconfigVM_Task, 'duration_secs': 0.355553} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.862749] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Reconfigured VM instance instance-00000042 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 902.869926] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b644c872-f9a0-4315-9cf4-bc99a6b45259 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.886470] env[61857]: DEBUG oslo_vmware.api [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 902.886470] env[61857]: value = "task-951214" [ 902.886470] env[61857]: _type = "Task" [ 902.886470] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.898451] env[61857]: DEBUG oslo_vmware.api [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951214, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.899893] env[61857]: DEBUG nova.policy [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac8387df3290404283263440672de653', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3bbcf7c2937040e1906e2273a07b671b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 903.090791] env[61857]: INFO nova.compute.manager [-] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Took 1.52 seconds to deallocate network for instance. [ 903.174475] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b07c98-ec34-41c6-be9e-d22469453421 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.184232] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.184232] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.186201] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc051537-cd57-418d-bcf5-78dafbfc5404 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.218999] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b82d303-0595-49b3-b84f-1cfccc722c0f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.232262] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39c2048-d827-4922-b9ce-19acccecbcd9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.248855] env[61857]: DEBUG nova.compute.provider_tree [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.300257] env[61857]: DEBUG nova.compute.manager [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 903.351789] env[61857]: DEBUG nova.network.neutron [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Successfully created port: e4c39a04-24d7-48d1-a089-2304778bd059 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 903.400266] env[61857]: DEBUG oslo_vmware.api [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951214, 'name': ReconfigVM_Task, 'duration_secs': 0.195876} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.400612] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214149', 'volume_id': '53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'name': 'volume-53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '16b447b5-426a-4478-9d44-ae32b41dee50', 'attached_at': '', 'detached_at': '', 'volume_id': '53af836e-3d73-4b7e-86c0-3d1641d68f9b', 'serial': '53af836e-3d73-4b7e-86c0-3d1641d68f9b'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 903.487859] env[61857]: DEBUG nova.compute.manager [req-c0378ad1-782a-4275-8a4c-266f7d4435bb req-d0a910c3-abfe-4ca1-a87b-41e7cffe7a2e service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Received event network-changed-639da037-e9cb-4624-994d-e7faab16e538 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 903.487973] env[61857]: DEBUG nova.compute.manager [req-c0378ad1-782a-4275-8a4c-266f7d4435bb req-d0a910c3-abfe-4ca1-a87b-41e7cffe7a2e service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Refreshing instance network info cache due to event network-changed-639da037-e9cb-4624-994d-e7faab16e538. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 903.488253] env[61857]: DEBUG oslo_concurrency.lockutils [req-c0378ad1-782a-4275-8a4c-266f7d4435bb req-d0a910c3-abfe-4ca1-a87b-41e7cffe7a2e service nova] Acquiring lock "refresh_cache-0de8e67e-24b9-48d0-ac90-b5ec5b93e243" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.488425] env[61857]: DEBUG oslo_concurrency.lockutils [req-c0378ad1-782a-4275-8a4c-266f7d4435bb req-d0a910c3-abfe-4ca1-a87b-41e7cffe7a2e service nova] Acquired lock "refresh_cache-0de8e67e-24b9-48d0-ac90-b5ec5b93e243" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.488775] env[61857]: DEBUG nova.network.neutron [req-c0378ad1-782a-4275-8a4c-266f7d4435bb req-d0a910c3-abfe-4ca1-a87b-41e7cffe7a2e service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Refreshing network info cache for port 639da037-e9cb-4624-994d-e7faab16e538 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 903.601127] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.690500] env[61857]: DEBUG nova.compute.manager [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 903.753967] env[61857]: DEBUG nova.scheduler.client.report [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.758670] env[61857]: DEBUG nova.network.neutron [-] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.958955] env[61857]: DEBUG nova.objects.instance [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lazy-loading 'flavor' on Instance uuid 16b447b5-426a-4478-9d44-ae32b41dee50 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.218540] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.258154] env[61857]: DEBUG nova.compute.manager [req-2d6215a5-5f2a-4469-9aef-1f00eda47235 req-baeab263-79c7-49ef-ab07-af852c5172c7 service nova] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Received event network-vif-deleted-03542c83-4268-49c5-a829-9d7750f2a70a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.261398] env[61857]: DEBUG oslo_concurrency.lockutils [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.979s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.264927] env[61857]: DEBUG oslo_concurrency.lockutils [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.473s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.264927] env[61857]: DEBUG oslo_concurrency.lockutils [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.266231] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.645s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.266594] env[61857]: DEBUG nova.objects.instance [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lazy-loading 'resources' on Instance uuid ffa63de7-dd32-4908-92c5-755b3c0799ef {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.268117] env[61857]: INFO nova.compute.manager [-] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Took 1.56 seconds to deallocate network for instance. [ 904.295587] env[61857]: INFO nova.scheduler.client.report [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Deleted allocations for instance 6957dc98-8c60-4fdd-83dd-be2f13825c6d [ 904.297718] env[61857]: INFO nova.scheduler.client.report [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleted allocations for instance 30d905ed-831d-44ae-807c-062de9a7e9fb [ 904.310617] env[61857]: DEBUG nova.compute.manager [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 904.342913] env[61857]: DEBUG nova.network.neutron [req-c0378ad1-782a-4275-8a4c-266f7d4435bb req-d0a910c3-abfe-4ca1-a87b-41e7cffe7a2e service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Updated VIF entry in instance network info cache for port 639da037-e9cb-4624-994d-e7faab16e538. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 904.343322] env[61857]: DEBUG nova.network.neutron [req-c0378ad1-782a-4275-8a4c-266f7d4435bb req-d0a910c3-abfe-4ca1-a87b-41e7cffe7a2e service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Updating instance_info_cache with network_info: [{"id": "639da037-e9cb-4624-994d-e7faab16e538", "address": "fa:16:3e:4a:86:b0", "network": {"id": "aa2712a7-cfd9-4139-8b6d-508cc240cee4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-131420680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ff67efd6404381a52a4001ec774a72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfbfc55d-8126-40dd-998e-8600ea92f97c", "external-id": "nsx-vlan-transportzone-650", "segmentation_id": 650, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap639da037-e9", "ovs_interfaceid": "639da037-e9cb-4624-994d-e7faab16e538", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.347374] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.347617] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.347863] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.348146] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.348355] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.348522] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.348738] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.348902] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.349086] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.349263] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.349445] env[61857]: DEBUG nova.virt.hardware [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.350398] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0023542f-4bc9-4c5c-9637-8c860d45c546 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.361160] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b9b004-4300-4679-abb2-d8940b592992 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.777169] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.814938] env[61857]: DEBUG oslo_concurrency.lockutils [None req-717d701d-54df-473e-a0a6-f382a8f290f8 tempest-AttachInterfacesV270Test-714469640 tempest-AttachInterfacesV270Test-714469640-project-member] Lock "6957dc98-8c60-4fdd-83dd-be2f13825c6d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.873s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.816128] env[61857]: DEBUG oslo_concurrency.lockutils [None req-791c9ae0-a9ff-40e1-b8c3-6a13cbea2630 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "30d905ed-831d-44ae-807c-062de9a7e9fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.152s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.854220] env[61857]: DEBUG oslo_concurrency.lockutils [req-c0378ad1-782a-4275-8a4c-266f7d4435bb req-d0a910c3-abfe-4ca1-a87b-41e7cffe7a2e service nova] Releasing lock "refresh_cache-0de8e67e-24b9-48d0-ac90-b5ec5b93e243" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.924998] env[61857]: DEBUG nova.network.neutron [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Successfully updated port: e4c39a04-24d7-48d1-a089-2304778bd059 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.973121] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b78bb7a5-d826-49e7-ac41-01bd6e430f2e tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.312s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.999096] env[61857]: DEBUG oslo_concurrency.lockutils [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "16b447b5-426a-4478-9d44-ae32b41dee50" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.999699] env[61857]: DEBUG oslo_concurrency.lockutils [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.073427] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8566ec-fcb5-4d7b-8a9a-1afa310bcdda {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.082597] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22415b8d-535f-4520-90ba-30ff8f7f5937 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.114374] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3746576e-2d3b-4f5d-a47e-57604bebcd26 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.122895] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b623f620-c456-4f6d-81af-65cd09fc6edd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.137397] env[61857]: DEBUG nova.compute.provider_tree [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.428814] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.428814] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.428814] env[61857]: DEBUG nova.network.neutron [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 905.504026] env[61857]: INFO nova.compute.manager [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Detaching volume b8ef1d4a-6429-449a-942d-0ea7be737467 [ 905.537210] env[61857]: INFO nova.virt.block_device [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Attempting to driver detach volume b8ef1d4a-6429-449a-942d-0ea7be737467 from mountpoint /dev/sdc [ 905.537210] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Volume detach. Driver type: vmdk {{(pid=61857) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 905.537210] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214153', 'volume_id': 'b8ef1d4a-6429-449a-942d-0ea7be737467', 'name': 'volume-b8ef1d4a-6429-449a-942d-0ea7be737467', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '16b447b5-426a-4478-9d44-ae32b41dee50', 'attached_at': '', 'detached_at': '', 'volume_id': 'b8ef1d4a-6429-449a-942d-0ea7be737467', 'serial': 'b8ef1d4a-6429-449a-942d-0ea7be737467'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 905.537210] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b27e21-bc74-4c45-bb5d-37e301e5c217 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.571035] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540a7268-6a45-4a56-b844-fd3a23fb13de {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.582905] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f08f6dd-d85e-4307-b532-6f293df5756a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.605124] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846cd593-61c8-405a-9062-aedf7e6410ea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.623760] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] The volume has not been displaced from its original location: [datastore2] volume-b8ef1d4a-6429-449a-942d-0ea7be737467/volume-b8ef1d4a-6429-449a-942d-0ea7be737467.vmdk. No consolidation needed. {{(pid=61857) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 905.629633] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Reconfiguring VM instance instance-00000042 to detach disk 2002 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 905.630112] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c90cabd-34ad-499b-944f-3fb255438df6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.644497] env[61857]: DEBUG nova.scheduler.client.report [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.657710] env[61857]: DEBUG oslo_vmware.api [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 905.657710] env[61857]: value = "task-951216" [ 905.657710] env[61857]: _type = "Task" [ 905.657710] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.667683] env[61857]: DEBUG oslo_vmware.api [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951216, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.827410] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Volume attach. Driver type: vmdk {{(pid=61857) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 905.827733] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214156', 'volume_id': '9647e67e-e4e4-4753-88a6-c144896c3f9c', 'name': 'volume-9647e67e-e4e4-4753-88a6-c144896c3f9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd979a8a1-a5b5-428e-9b60-7d827337218f', 'attached_at': '', 'detached_at': '', 'volume_id': '9647e67e-e4e4-4753-88a6-c144896c3f9c', 'serial': '9647e67e-e4e4-4753-88a6-c144896c3f9c'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 905.828655] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8c887d-4186-4da6-a858-23f4a4fe6c4d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.847224] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1880a8-bd7f-416d-bb75-e3378c891e1c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.878424] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] volume-9647e67e-e4e4-4753-88a6-c144896c3f9c/volume-9647e67e-e4e4-4753-88a6-c144896c3f9c.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.878801] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bacd901e-22a5-4509-b573-4969a7f24842 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.900583] env[61857]: DEBUG oslo_vmware.api [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 905.900583] env[61857]: value = "task-951217" [ 905.900583] env[61857]: _type = "Task" [ 905.900583] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.909434] env[61857]: DEBUG oslo_vmware.api [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951217, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.992771] env[61857]: DEBUG nova.network.neutron [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 906.151043] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.153846] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.447s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.154635] env[61857]: DEBUG nova.objects.instance [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lazy-loading 'resources' on Instance uuid 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.170702] env[61857]: DEBUG oslo_vmware.api [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951216, 'name': ReconfigVM_Task, 'duration_secs': 0.218466} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.173569] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Reconfigured VM instance instance-00000042 to detach disk 2002 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 906.180103] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f353e25e-273a-4c01-b7ba-fe916f01c776 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.192821] env[61857]: INFO nova.scheduler.client.report [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Deleted allocations for instance ffa63de7-dd32-4908-92c5-755b3c0799ef [ 906.205232] env[61857]: DEBUG oslo_vmware.api [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 906.205232] env[61857]: value = "task-951218" [ 906.205232] env[61857]: _type = "Task" [ 906.205232] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.213951] env[61857]: DEBUG oslo_vmware.api [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951218, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.280656] env[61857]: DEBUG nova.network.neutron [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance_info_cache with network_info: [{"id": "e4c39a04-24d7-48d1-a089-2304778bd059", "address": "fa:16:3e:45:7a:cc", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4c39a04-24", "ovs_interfaceid": "e4c39a04-24d7-48d1-a089-2304778bd059", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.288568] env[61857]: DEBUG nova.compute.manager [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Received event network-vif-plugged-e4c39a04-24d7-48d1-a089-2304778bd059 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 906.288821] env[61857]: DEBUG oslo_concurrency.lockutils [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] Acquiring lock "18a632be-322f-48a9-becf-51ff3b735ad4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.289185] env[61857]: DEBUG oslo_concurrency.lockutils [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] Lock "18a632be-322f-48a9-becf-51ff3b735ad4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.289421] env[61857]: DEBUG oslo_concurrency.lockutils [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] Lock "18a632be-322f-48a9-becf-51ff3b735ad4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.289611] env[61857]: DEBUG nova.compute.manager [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] No waiting events found dispatching network-vif-plugged-e4c39a04-24d7-48d1-a089-2304778bd059 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.289876] env[61857]: WARNING nova.compute.manager [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Received unexpected event network-vif-plugged-e4c39a04-24d7-48d1-a089-2304778bd059 for instance with vm_state building and task_state spawning. [ 906.289972] env[61857]: DEBUG nova.compute.manager [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Received event network-changed-e4c39a04-24d7-48d1-a089-2304778bd059 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 906.290108] env[61857]: DEBUG nova.compute.manager [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Refreshing instance network info cache due to event network-changed-e4c39a04-24d7-48d1-a089-2304778bd059. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 906.290335] env[61857]: DEBUG oslo_concurrency.lockutils [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] Acquiring lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.410968] env[61857]: DEBUG oslo_vmware.api [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951217, 'name': ReconfigVM_Task, 'duration_secs': 0.361232} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.411323] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Reconfigured VM instance instance-00000050 to attach disk [datastore1] volume-9647e67e-e4e4-4753-88a6-c144896c3f9c/volume-9647e67e-e4e4-4753-88a6-c144896c3f9c.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.415953] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8763413-2acb-49f2-b01a-97afc2f786c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.436098] env[61857]: DEBUG oslo_vmware.api [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 906.436098] env[61857]: value = "task-951219" [ 906.436098] env[61857]: _type = "Task" [ 906.436098] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.445561] env[61857]: DEBUG oslo_vmware.api [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951219, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.704240] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5876b299-4f2c-4a02-99a8-40806939fef0 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.358s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.705398] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 19.174s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.705625] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "ffa63de7-dd32-4908-92c5-755b3c0799ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.705832] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.706044] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.711045] env[61857]: INFO nova.compute.manager [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Terminating instance [ 906.712656] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.712878] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquired lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.713087] env[61857]: DEBUG nova.network.neutron [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 906.720308] env[61857]: DEBUG oslo_vmware.api [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951218, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.784036] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.784421] env[61857]: DEBUG nova.compute.manager [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Instance network_info: |[{"id": "e4c39a04-24d7-48d1-a089-2304778bd059", "address": "fa:16:3e:45:7a:cc", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4c39a04-24", "ovs_interfaceid": "e4c39a04-24d7-48d1-a089-2304778bd059", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 906.784751] env[61857]: DEBUG oslo_concurrency.lockutils [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] Acquired lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.784942] env[61857]: DEBUG nova.network.neutron [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Refreshing network info cache for port e4c39a04-24d7-48d1-a089-2304778bd059 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 906.786165] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:7a:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3a80436-f7a9-431a-acec-aca3d76e3f9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4c39a04-24d7-48d1-a089-2304778bd059', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.803713] env[61857]: DEBUG oslo.service.loopingcall [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.811785] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 906.812483] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75435e81-d50e-413f-a7c7-10a802dba4e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.841953] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.841953] env[61857]: value = "task-951220" [ 906.841953] env[61857]: _type = "Task" [ 906.841953] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.853884] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951220, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.950747] env[61857]: DEBUG oslo_vmware.api [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951219, 'name': ReconfigVM_Task, 'duration_secs': 0.153512} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.951295] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214156', 'volume_id': '9647e67e-e4e4-4753-88a6-c144896c3f9c', 'name': 'volume-9647e67e-e4e4-4753-88a6-c144896c3f9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd979a8a1-a5b5-428e-9b60-7d827337218f', 'attached_at': '', 'detached_at': '', 'volume_id': '9647e67e-e4e4-4753-88a6-c144896c3f9c', 'serial': '9647e67e-e4e4-4753-88a6-c144896c3f9c'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 907.007134] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb526190-44c3-46d4-bbfe-0c3f7ebd51ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.017031] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b553f352-1c55-4bbd-af5d-43ac231c1867 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.058944] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2053be5-dff5-476a-857d-b75aef43c403 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.068548] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03feec22-0361-4ab2-9692-c92a3326470e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.084103] env[61857]: DEBUG nova.compute.provider_tree [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.215823] env[61857]: DEBUG nova.compute.utils [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Can not refresh info_cache because instance was not found {{(pid=61857) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 907.217845] env[61857]: DEBUG oslo_vmware.api [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951218, 'name': ReconfigVM_Task, 'duration_secs': 0.763586} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.218183] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214153', 'volume_id': 'b8ef1d4a-6429-449a-942d-0ea7be737467', 'name': 'volume-b8ef1d4a-6429-449a-942d-0ea7be737467', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '16b447b5-426a-4478-9d44-ae32b41dee50', 'attached_at': '', 'detached_at': '', 'volume_id': 'b8ef1d4a-6429-449a-942d-0ea7be737467', 'serial': 'b8ef1d4a-6429-449a-942d-0ea7be737467'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 907.249474] env[61857]: DEBUG nova.network.neutron [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 907.287585] env[61857]: DEBUG nova.network.neutron [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updated VIF entry in instance network info cache for port e4c39a04-24d7-48d1-a089-2304778bd059. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 907.287962] env[61857]: DEBUG nova.network.neutron [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance_info_cache with network_info: [{"id": "e4c39a04-24d7-48d1-a089-2304778bd059", "address": "fa:16:3e:45:7a:cc", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4c39a04-24", "ovs_interfaceid": "e4c39a04-24d7-48d1-a089-2304778bd059", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.358134] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951220, 'name': CreateVM_Task, 'duration_secs': 0.351545} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.358412] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 907.359148] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.359339] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.359686] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 907.359958] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b64eef2-2893-4876-852f-e5422eabb86d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.366648] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 907.366648] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52df2d64-7515-6484-a387-ef1c1346e33c" [ 907.366648] env[61857]: _type = "Task" [ 907.366648] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.382597] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52df2d64-7515-6484-a387-ef1c1346e33c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.399293] env[61857]: DEBUG nova.network.neutron [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.587213] env[61857]: DEBUG nova.scheduler.client.report [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.616544] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.616876] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.617119] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.617354] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.617502] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.620319] env[61857]: INFO nova.compute.manager [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Terminating instance [ 907.622579] env[61857]: DEBUG nova.compute.manager [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 907.622802] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 907.623663] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3366a71c-df0f-4a2c-8eda-9a947c6c1a2e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.631820] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 907.632093] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed381b37-6264-4a40-aa43-028421b94367 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.640165] env[61857]: DEBUG oslo_vmware.api [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 907.640165] env[61857]: value = "task-951221" [ 907.640165] env[61857]: _type = "Task" [ 907.640165] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.652305] env[61857]: DEBUG oslo_vmware.api [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951221, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.763255] env[61857]: DEBUG nova.objects.instance [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lazy-loading 'flavor' on Instance uuid 16b447b5-426a-4478-9d44-ae32b41dee50 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.790740] env[61857]: DEBUG oslo_concurrency.lockutils [req-293e5210-ea7d-4c37-998e-c7e5b7bfa1d6 req-3691bec0-8adc-4413-92be-6716806dfda5 service nova] Releasing lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.881176] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52df2d64-7515-6484-a387-ef1c1346e33c, 'name': SearchDatastore_Task, 'duration_secs': 0.010399} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.881176] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.881176] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.881176] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.881176] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.881176] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.881176] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fdd202a1-41a9-4b1f-9d39-7873bf89a7f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.889779] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.889967] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 907.890724] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a8e79e7-4c98-4869-8652-e750c1c76e42 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.896348] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 907.896348] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524bbad9-2de3-fbb4-8a8d-264976613e6d" [ 907.896348] env[61857]: _type = "Task" [ 907.896348] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.907036] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Releasing lock "refresh_cache-ffa63de7-dd32-4908-92c5-755b3c0799ef" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.907036] env[61857]: DEBUG nova.compute.manager [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 907.907036] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 907.907036] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524bbad9-2de3-fbb4-8a8d-264976613e6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.907036] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17a24c6e-4597-48cf-8bb4-5188b0b8bad3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.915864] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bcf07c5-a677-4357-a9c0-57688ae6d602 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.948378] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ffa63de7-dd32-4908-92c5-755b3c0799ef could not be found. [ 907.948626] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 907.948821] env[61857]: INFO nova.compute.manager [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Took 0.04 seconds to destroy the instance on the hypervisor. [ 907.949154] env[61857]: DEBUG oslo.service.loopingcall [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.949362] env[61857]: DEBUG nova.compute.manager [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.949461] env[61857]: DEBUG nova.network.neutron [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 908.009120] env[61857]: DEBUG nova.objects.instance [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lazy-loading 'flavor' on Instance uuid d979a8a1-a5b5-428e-9b60-7d827337218f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.092119] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.938s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.096564] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.904s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.096564] env[61857]: DEBUG nova.objects.instance [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lazy-loading 'resources' on Instance uuid 7a9252da-2584-40ed-9d28-ca7341ed5165 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.114056] env[61857]: INFO nova.scheduler.client.report [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Deleted allocations for instance 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc [ 908.152976] env[61857]: DEBUG oslo_vmware.api [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951221, 'name': PowerOffVM_Task, 'duration_secs': 0.201776} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.153336] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 908.153515] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 908.153881] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc4dfe28-8327-4016-ba03-e83a1a33cfaf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.195412] env[61857]: DEBUG nova.network.neutron [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 908.226506] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 908.226830] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 908.227496] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Deleting the datastore file [datastore1] 24d3d515-e03a-4b4e-bb8e-bc18537125ac {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 908.227496] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-758fab1b-1d07-47be-b87c-ca6f2be8c749 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.235976] env[61857]: DEBUG oslo_vmware.api [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for the task: (returnval){ [ 908.235976] env[61857]: value = "task-951223" [ 908.235976] env[61857]: _type = "Task" [ 908.235976] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.245165] env[61857]: DEBUG oslo_vmware.api [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951223, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.413701] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524bbad9-2de3-fbb4-8a8d-264976613e6d, 'name': SearchDatastore_Task, 'duration_secs': 0.008614} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.414814] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66c3611f-cdeb-46ef-b7bd-80b2c1758ec6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.423099] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 908.423099] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52f4436b-52d0-ef07-fa02-324c3092a09a" [ 908.423099] env[61857]: _type = "Task" [ 908.423099] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.432519] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f4436b-52d0-ef07-fa02-324c3092a09a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.519121] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0d5e02cd-cf11-43ef-9d05-b44b2f31f529 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.311s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.623441] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f263ec77-a415-4898-9a9f-9d1d56a2515c tempest-ServerAddressesTestJSON-974761464 tempest-ServerAddressesTestJSON-974761464-project-member] Lock "0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.981s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.678275] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 908.679068] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 908.699928] env[61857]: DEBUG nova.network.neutron [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.752018] env[61857]: DEBUG oslo_vmware.api [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Task: {'id': task-951223, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126627} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.753274] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 908.753465] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 908.753748] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 908.753848] env[61857]: INFO nova.compute.manager [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Took 1.13 seconds to destroy the instance on the hypervisor. [ 908.754130] env[61857]: DEBUG oslo.service.loopingcall [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.755587] env[61857]: DEBUG nova.compute.manager [-] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 908.755700] env[61857]: DEBUG nova.network.neutron [-] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 908.771739] env[61857]: DEBUG oslo_concurrency.lockutils [None req-445a2a25-5602-4c9a-9b3d-cd2f6d025453 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.771s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.914240] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77e42e5-d46e-4099-81c8-79c4b1e6afea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.928552] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.929148] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.937068] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2279cac0-f8fb-4c79-b254-2e27506fc289 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.955031] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f4436b-52d0-ef07-fa02-324c3092a09a, 'name': SearchDatastore_Task, 'duration_secs': 0.011494} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.983458] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.983701] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 18a632be-322f-48a9-becf-51ff3b735ad4/18a632be-322f-48a9-becf-51ff3b735ad4.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 908.984785] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05d19992-9a42-4929-95bd-e3d15cc6b33d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.987605] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371d2bf4-223e-4dc5-b711-b52ab3616994 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.990687] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "d979a8a1-a5b5-428e-9b60-7d827337218f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.991323] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.991323] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "d979a8a1-a5b5-428e-9b60-7d827337218f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.991465] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.991685] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.995133] env[61857]: INFO nova.compute.manager [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Terminating instance [ 908.999769] env[61857]: DEBUG nova.compute.manager [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.000009] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 909.001438] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ddff6b2-37f8-441b-9ef8-2fb04af9ebff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.005793] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72824868-5d50-49ec-9975-d49752cd4b02 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.009609] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 909.009609] env[61857]: value = "task-951224" [ 909.009609] env[61857]: _type = "Task" [ 909.009609] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.025104] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 909.025104] env[61857]: value = "task-951225" [ 909.025104] env[61857]: _type = "Task" [ 909.025104] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.025885] env[61857]: DEBUG nova.compute.provider_tree [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.038365] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951224, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.046654] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951225, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.188919] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 909.188919] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 909.204678] env[61857]: INFO nova.compute.manager [-] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Took 1.26 seconds to deallocate network for instance. [ 909.374468] env[61857]: DEBUG nova.compute.manager [req-7fa78dee-99ac-42ec-bf3c-4c2ff6f598cc req-8469e348-1747-42b5-bf3a-e59cd511d523 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Received event network-vif-deleted-25a85668-66ff-4cf6-8cef-68b4d1a2e828 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.374748] env[61857]: INFO nova.compute.manager [req-7fa78dee-99ac-42ec-bf3c-4c2ff6f598cc req-8469e348-1747-42b5-bf3a-e59cd511d523 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Neutron deleted interface 25a85668-66ff-4cf6-8cef-68b4d1a2e828; detaching it from the instance and deleting it from the info cache [ 909.374933] env[61857]: DEBUG nova.network.neutron [req-7fa78dee-99ac-42ec-bf3c-4c2ff6f598cc req-8469e348-1747-42b5-bf3a-e59cd511d523 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.432458] env[61857]: DEBUG nova.compute.manager [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 909.521275] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951224, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467403} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.521734] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 18a632be-322f-48a9-becf-51ff3b735ad4/18a632be-322f-48a9-becf-51ff3b735ad4.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 909.521895] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.522041] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a4badc7d-dfb3-4b79-8d90-6a599da10afb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.530273] env[61857]: DEBUG nova.scheduler.client.report [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.533638] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 909.533638] env[61857]: value = "task-951226" [ 909.533638] env[61857]: _type = "Task" [ 909.533638] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.546895] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951226, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.550668] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951225, 'name': PowerOffVM_Task, 'duration_secs': 0.240549} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.550958] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 909.551226] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Volume detach. Driver type: vmdk {{(pid=61857) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 909.551464] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214156', 'volume_id': '9647e67e-e4e4-4753-88a6-c144896c3f9c', 'name': 'volume-9647e67e-e4e4-4753-88a6-c144896c3f9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd979a8a1-a5b5-428e-9b60-7d827337218f', 'attached_at': '', 'detached_at': '', 'volume_id': '9647e67e-e4e4-4753-88a6-c144896c3f9c', 'serial': '9647e67e-e4e4-4753-88a6-c144896c3f9c'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 909.552520] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e14d02e6-84dc-4aa1-ac85-2ab4b0a05b80 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.575948] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b720fe-e821-4c8d-966d-eddef414e887 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.584589] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb66981-47ca-4ff9-8ff2-eee7091f5d5e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.607933] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e2c817-1b6f-45a9-81fd-adefd6a5439e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.625185] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] The volume has not been displaced from its original location: [datastore1] volume-9647e67e-e4e4-4753-88a6-c144896c3f9c/volume-9647e67e-e4e4-4753-88a6-c144896c3f9c.vmdk. No consolidation needed. {{(pid=61857) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 909.630233] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Reconfiguring VM instance instance-00000050 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 909.630594] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72f00b96-bdbf-4de1-9000-1f2fcb8981de {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.657948] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 909.657948] env[61857]: value = "task-951227" [ 909.657948] env[61857]: _type = "Task" [ 909.657948] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.669531] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951227, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.711584] env[61857]: INFO nova.compute.manager [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Instance disappeared during terminate [ 909.711845] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2824f70c-090c-49d1-b522-91bd4fbc048e tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "ffa63de7-dd32-4908-92c5-755b3c0799ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.006s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.723797] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.723938] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquired lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.724141] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Forcefully refreshing network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 909.786531] env[61857]: DEBUG nova.network.neutron [-] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.808772] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "16b447b5-426a-4478-9d44-ae32b41dee50" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.809067] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.809395] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "16b447b5-426a-4478-9d44-ae32b41dee50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.809640] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.810064] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.812066] env[61857]: INFO nova.compute.manager [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Terminating instance [ 909.813895] env[61857]: DEBUG nova.compute.manager [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.814111] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 909.815209] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5f2767-d73b-46df-806b-322558b8ac12 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.824679] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 909.824944] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a1c7b66f-45f6-4988-b9ce-ada5418ab8aa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.832942] env[61857]: DEBUG oslo_vmware.api [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 909.832942] env[61857]: value = "task-951228" [ 909.832942] env[61857]: _type = "Task" [ 909.832942] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.842710] env[61857]: DEBUG oslo_vmware.api [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951228, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.878386] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cdafa2ba-8550-4296-95c2-073bdd7de84a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.889132] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32d57ea-fdde-4907-956e-5218fc27f2dd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.930561] env[61857]: DEBUG nova.compute.manager [req-7fa78dee-99ac-42ec-bf3c-4c2ff6f598cc req-8469e348-1747-42b5-bf3a-e59cd511d523 service nova] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Detach interface failed, port_id=25a85668-66ff-4cf6-8cef-68b4d1a2e828, reason: Instance 24d3d515-e03a-4b4e-bb8e-bc18537125ac could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 909.962345] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.038646] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.942s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.039772] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.182s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.043999] env[61857]: INFO nova.compute.claims [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.060059] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951226, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089082} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.060428] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 910.061289] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402c460e-cfcc-48e5-a8ac-7644d011a829 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.090080] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 18a632be-322f-48a9-becf-51ff3b735ad4/18a632be-322f-48a9-becf-51ff3b735ad4.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 910.091510] env[61857]: INFO nova.scheduler.client.report [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Deleted allocations for instance 7a9252da-2584-40ed-9d28-ca7341ed5165 [ 910.092757] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35450f1c-bb45-4a85-bfa3-7c5d28794056 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.124532] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 910.124532] env[61857]: value = "task-951229" [ 910.124532] env[61857]: _type = "Task" [ 910.124532] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.139974] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951229, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.169868] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951227, 'name': ReconfigVM_Task, 'duration_secs': 0.375388} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.170216] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Reconfigured VM instance instance-00000050 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 910.176347] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6221a9b1-d7fe-4059-9592-6a7ad07b32ad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.194641] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 910.194641] env[61857]: value = "task-951230" [ 910.194641] env[61857]: _type = "Task" [ 910.194641] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.208850] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951230, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.249788] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 910.290042] env[61857]: INFO nova.compute.manager [-] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Took 1.53 seconds to deallocate network for instance. [ 910.348304] env[61857]: DEBUG oslo_vmware.api [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951228, 'name': PowerOffVM_Task, 'duration_secs': 0.271044} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.348598] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 910.348958] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 910.349091] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1396959-81aa-4f9c-8308-d3328d535e81 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.452185] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 910.452640] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 910.452640] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Deleting the datastore file [datastore2] 16b447b5-426a-4478-9d44-ae32b41dee50 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.452910] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a17c1903-2813-42e0-9c13-860483176db3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.460270] env[61857]: DEBUG oslo_vmware.api [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for the task: (returnval){ [ 910.460270] env[61857]: value = "task-951232" [ 910.460270] env[61857]: _type = "Task" [ 910.460270] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.469250] env[61857]: DEBUG oslo_vmware.api [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951232, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.620180] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de5ad970-183a-4900-b772-d66448fe69d1 tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "7a9252da-2584-40ed-9d28-ca7341ed5165" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.564s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.638343] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951229, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.705896] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951230, 'name': ReconfigVM_Task, 'duration_secs': 0.501172} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.706268] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214156', 'volume_id': '9647e67e-e4e4-4753-88a6-c144896c3f9c', 'name': 'volume-9647e67e-e4e4-4753-88a6-c144896c3f9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd979a8a1-a5b5-428e-9b60-7d827337218f', 'attached_at': '', 'detached_at': '', 'volume_id': '9647e67e-e4e4-4753-88a6-c144896c3f9c', 'serial': '9647e67e-e4e4-4753-88a6-c144896c3f9c'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 910.706609] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 910.707432] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56324330-b85a-4060-b288-fa834f1e4448 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.715946] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 910.716276] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb5aff15-7199-438d-a732-7905c960dfe4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.797378] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.809255] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 910.809255] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 910.809437] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleting the datastore file [datastore2] d979a8a1-a5b5-428e-9b60-7d827337218f {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.809760] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cd1cc148-afe3-4f0e-9476-f5829fb06736 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.819703] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 910.819703] env[61857]: value = "task-951234" [ 910.819703] env[61857]: _type = "Task" [ 910.819703] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.830621] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951234, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.884813] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.972885] env[61857]: DEBUG oslo_vmware.api [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Task: {'id': task-951232, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.279053} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.973199] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.973399] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 910.973586] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 910.973809] env[61857]: INFO nova.compute.manager [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Took 1.16 seconds to destroy the instance on the hypervisor. [ 910.974015] env[61857]: DEBUG oslo.service.loopingcall [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.974225] env[61857]: DEBUG nova.compute.manager [-] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.974321] env[61857]: DEBUG nova.network.neutron [-] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 911.147504] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951229, 'name': ReconfigVM_Task, 'duration_secs': 0.704018} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.152555] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 18a632be-322f-48a9-becf-51ff3b735ad4/18a632be-322f-48a9-becf-51ff3b735ad4.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.153922] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1081d69e-4967-4b5c-b989-221a82f1f12a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.166450] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 911.166450] env[61857]: value = "task-951235" [ 911.166450] env[61857]: _type = "Task" [ 911.166450] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.182806] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951235, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.332201] env[61857]: DEBUG oslo_vmware.api [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951234, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146531} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.334031] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.334031] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 911.334031] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 911.334031] env[61857]: INFO nova.compute.manager [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Took 2.33 seconds to destroy the instance on the hypervisor. [ 911.334031] env[61857]: DEBUG oslo.service.loopingcall [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.334031] env[61857]: DEBUG nova.compute.manager [-] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.334300] env[61857]: DEBUG nova.network.neutron [-] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 911.387482] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Releasing lock "refresh_cache-e88aed62-6561-4e29-b041-46b19cc5ef63" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.387714] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Updated the network info_cache for instance {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 911.387951] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.388258] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.388468] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.388671] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.388862] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.389027] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.389168] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 911.389319] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.411531] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61250754-298c-4ced-9694-9b3984fff423 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.426066] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e334d6fb-e87f-4064-84c2-24375724f685 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.460178] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b75c7d-d3fc-43c6-9317-a12a823f83af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.468966] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235eddde-7084-4e80-903d-672ce9298f1d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.484019] env[61857]: DEBUG nova.compute.provider_tree [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.677249] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951235, 'name': Rename_Task, 'duration_secs': 0.156911} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.677602] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 911.677726] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2565fe4d-3a91-490c-9b72-e93ba0d9f5d4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.685191] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 911.685191] env[61857]: value = "task-951236" [ 911.685191] env[61857]: _type = "Task" [ 911.685191] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.722453] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951236, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.896380] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.936558] env[61857]: DEBUG nova.compute.manager [req-afdd2a68-538e-4a78-83ae-e28c104dd1cc req-d91d807e-81ad-4dcb-9a17-398d1064f704 service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Received event network-vif-deleted-4ba4d151-0e84-4f6d-aee1-6843bae4fe64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.936558] env[61857]: INFO nova.compute.manager [req-afdd2a68-538e-4a78-83ae-e28c104dd1cc req-d91d807e-81ad-4dcb-9a17-398d1064f704 service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Neutron deleted interface 4ba4d151-0e84-4f6d-aee1-6843bae4fe64; detaching it from the instance and deleting it from the info cache [ 911.936558] env[61857]: DEBUG nova.network.neutron [req-afdd2a68-538e-4a78-83ae-e28c104dd1cc req-d91d807e-81ad-4dcb-9a17-398d1064f704 service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.987462] env[61857]: DEBUG nova.scheduler.client.report [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.196864] env[61857]: DEBUG oslo_vmware.api [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951236, 'name': PowerOnVM_Task, 'duration_secs': 0.461672} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.197241] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 912.197433] env[61857]: INFO nova.compute.manager [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Took 7.89 seconds to spawn the instance on the hypervisor. [ 912.197657] env[61857]: DEBUG nova.compute.manager [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.198549] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ad2c70-178b-49e6-a3f1-a6fec6f21a38 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.330689] env[61857]: DEBUG nova.network.neutron [-] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.371605] env[61857]: DEBUG nova.network.neutron [-] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.439236] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-434d04f8-54ab-4fad-98c0-8dd4717f9182 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.450892] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e666ac-0450-49c3-b8dc-70641df1ec3c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.466944] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquiring lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.467267] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.488265] env[61857]: DEBUG nova.compute.manager [req-afdd2a68-538e-4a78-83ae-e28c104dd1cc req-d91d807e-81ad-4dcb-9a17-398d1064f704 service nova] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Detach interface failed, port_id=4ba4d151-0e84-4f6d-aee1-6843bae4fe64, reason: Instance d979a8a1-a5b5-428e-9b60-7d827337218f could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 912.493896] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.493896] env[61857]: DEBUG nova.compute.manager [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 912.496827] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.549s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.497090] env[61857]: DEBUG nova.objects.instance [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lazy-loading 'resources' on Instance uuid ddd4c436-405f-49f7-8c9b-de3b71725f63 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.720749] env[61857]: INFO nova.compute.manager [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Took 29.37 seconds to build instance. [ 912.833698] env[61857]: INFO nova.compute.manager [-] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Took 1.50 seconds to deallocate network for instance. [ 912.875474] env[61857]: INFO nova.compute.manager [-] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Took 1.90 seconds to deallocate network for instance. [ 912.929824] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "781a2790-e317-46fe-9be8-40c9e5f1f771" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.930167] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.930400] env[61857]: INFO nova.compute.manager [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Shelving [ 912.971098] env[61857]: DEBUG nova.compute.manager [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 912.999431] env[61857]: DEBUG nova.compute.utils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.000932] env[61857]: DEBUG nova.compute.manager [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 913.004525] env[61857]: DEBUG nova.network.neutron [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 913.086639] env[61857]: DEBUG nova.policy [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2ef7e845007475a8a19cf3f520c3f51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1acf476aabc4166bc8505a3442367c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.222544] env[61857]: DEBUG oslo_concurrency.lockutils [None req-90771401-fac4-4a0e-88b3-29bb61391e73 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "18a632be-322f-48a9-becf-51ff3b735ad4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.890s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.340953] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1dafaee-46c2-4d7e-9675-096baf65121c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.354578] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d6427f-bc4e-4d5a-a8f1-54e6d2322d08 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.390109] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.391328] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f336b2-16d8-476b-a999-87c5ccd1e444 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.396937] env[61857]: INFO nova.compute.manager [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Took 0.56 seconds to detach 1 volumes for instance. [ 913.405162] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb6797b-f7d5-4e3b-9f7d-1afb0204a241 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.424570] env[61857]: DEBUG nova.compute.provider_tree [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.430282] env[61857]: DEBUG nova.network.neutron [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Successfully created port: a703ce6c-14a3-4a8e-aa09-8714098d603e {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 913.439141] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.439385] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-189d02dc-7b38-4fa4-ba59-a7f3373bc20e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.449358] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 913.449358] env[61857]: value = "task-951237" [ 913.449358] env[61857]: _type = "Task" [ 913.449358] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.458659] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951237, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.495811] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.510973] env[61857]: DEBUG nova.compute.manager [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 913.914142] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.927592] env[61857]: DEBUG nova.scheduler.client.report [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.959795] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951237, 'name': PowerOffVM_Task, 'duration_secs': 0.210568} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.960097] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 913.960910] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c96173f-f718-4859-9a3b-14aae09bb29d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.965799] env[61857]: DEBUG nova.compute.manager [req-41cb02ae-1f29-45f1-a33d-50b8a79e1004 req-9d430528-f845-4b77-9e1f-e67ab3d0224f service nova] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Received event network-vif-deleted-71f818ca-5a91-42ed-8825-5141bf75a30e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.984909] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22d5a6e-22d3-4326-b658-8a33a88d0656 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.279275] env[61857]: DEBUG nova.compute.manager [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Stashing vm_state: active {{(pid=61857) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 914.432547] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.936s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.434917] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.853s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.435188] env[61857]: DEBUG nova.objects.instance [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lazy-loading 'resources' on Instance uuid 42d38f5e-628a-4030-85e1-3ec0595cf3c8 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.457551] env[61857]: INFO nova.scheduler.client.report [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Deleted allocations for instance ddd4c436-405f-49f7-8c9b-de3b71725f63 [ 914.498159] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 914.498497] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d5e0a615-b4ce-4c99-9f7f-36009d2a034e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.509682] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 914.509682] env[61857]: value = "task-951238" [ 914.509682] env[61857]: _type = "Task" [ 914.509682] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.521014] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951238, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.522183] env[61857]: DEBUG nova.compute.manager [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.561580] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.561875] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.562037] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.562222] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.562383] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.562622] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.562771] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.562937] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.563201] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.563554] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.563637] env[61857]: DEBUG nova.virt.hardware [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.564439] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b24d19f-d469-41bd-ae58-5c12fef3632f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.573449] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43d22af-a312-4d91-8b3e-4a95555cae08 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.796059] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.967067] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4f80f511-c412-444a-b6f0-d8e3c8cc1d6b tempest-ListServersNegativeTestJSON-1694063133 tempest-ListServersNegativeTestJSON-1694063133-project-member] Lock "ddd4c436-405f-49f7-8c9b-de3b71725f63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.972s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.023761] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951238, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.183573] env[61857]: DEBUG nova.network.neutron [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Successfully updated port: a703ce6c-14a3-4a8e-aa09-8714098d603e {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.247037] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09210cc9-21ae-4a84-a29d-81a9ead87fc9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.256031] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b265d6-87fa-4d64-bb39-4d69de2bbd99 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.287925] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e7b9ee-eec8-44ae-a7c0-c7ece3d32b63 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.296402] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c00a19ad-1084-4101-afe2-857002f74682 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.310552] env[61857]: DEBUG nova.compute.provider_tree [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.523020] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951238, 'name': CreateSnapshot_Task, 'duration_secs': 0.731913} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.523336] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 915.524162] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4953130c-3452-4fec-8b1f-a058201f31ff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.685736] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "refresh_cache-992bcab3-7396-495f-96a1-e99ecd13c961" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.685886] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "refresh_cache-992bcab3-7396-495f-96a1-e99ecd13c961" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.686053] env[61857]: DEBUG nova.network.neutron [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.816578] env[61857]: DEBUG nova.scheduler.client.report [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.992878] env[61857]: DEBUG nova.compute.manager [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Received event network-vif-plugged-a703ce6c-14a3-4a8e-aa09-8714098d603e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 915.993230] env[61857]: DEBUG oslo_concurrency.lockutils [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] Acquiring lock "992bcab3-7396-495f-96a1-e99ecd13c961-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.993336] env[61857]: DEBUG oslo_concurrency.lockutils [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] Lock "992bcab3-7396-495f-96a1-e99ecd13c961-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.993512] env[61857]: DEBUG oslo_concurrency.lockutils [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] Lock "992bcab3-7396-495f-96a1-e99ecd13c961-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.993691] env[61857]: DEBUG nova.compute.manager [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] No waiting events found dispatching network-vif-plugged-a703ce6c-14a3-4a8e-aa09-8714098d603e {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 915.993867] env[61857]: WARNING nova.compute.manager [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Received unexpected event network-vif-plugged-a703ce6c-14a3-4a8e-aa09-8714098d603e for instance with vm_state building and task_state spawning. [ 915.994046] env[61857]: DEBUG nova.compute.manager [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Received event network-changed-a703ce6c-14a3-4a8e-aa09-8714098d603e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 915.994237] env[61857]: DEBUG nova.compute.manager [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Refreshing instance network info cache due to event network-changed-a703ce6c-14a3-4a8e-aa09-8714098d603e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 915.994421] env[61857]: DEBUG oslo_concurrency.lockutils [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] Acquiring lock "refresh_cache-992bcab3-7396-495f-96a1-e99ecd13c961" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.046990] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 916.047360] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f84ba96d-93de-4204-ae84-d5fda7acad27 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.057741] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 916.057741] env[61857]: value = "task-951239" [ 916.057741] env[61857]: _type = "Task" [ 916.057741] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.067128] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951239, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.222351] env[61857]: DEBUG nova.network.neutron [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 916.322443] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.887s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.324979] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.725s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.325272] env[61857]: DEBUG nova.objects.instance [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lazy-loading 'resources' on Instance uuid 33cb5bbc-926d-42ee-b483-8d1e24707e40 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.345433] env[61857]: INFO nova.scheduler.client.report [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Deleted allocations for instance 42d38f5e-628a-4030-85e1-3ec0595cf3c8 [ 916.366463] env[61857]: DEBUG nova.network.neutron [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Updating instance_info_cache with network_info: [{"id": "a703ce6c-14a3-4a8e-aa09-8714098d603e", "address": "fa:16:3e:d1:82:65", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa703ce6c-14", "ovs_interfaceid": "a703ce6c-14a3-4a8e-aa09-8714098d603e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.570442] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951239, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.855162] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2f0ad48d-bfee-4f46-970f-1a84be0e5f90 tempest-ListImageFiltersTestJSON-987379334 tempest-ListImageFiltersTestJSON-987379334-project-member] Lock "42d38f5e-628a-4030-85e1-3ec0595cf3c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.289s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.875055] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "refresh_cache-992bcab3-7396-495f-96a1-e99ecd13c961" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.875055] env[61857]: DEBUG nova.compute.manager [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Instance network_info: |[{"id": "a703ce6c-14a3-4a8e-aa09-8714098d603e", "address": "fa:16:3e:d1:82:65", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa703ce6c-14", "ovs_interfaceid": "a703ce6c-14a3-4a8e-aa09-8714098d603e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 916.875055] env[61857]: DEBUG oslo_concurrency.lockutils [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] Acquired lock "refresh_cache-992bcab3-7396-495f-96a1-e99ecd13c961" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.875055] env[61857]: DEBUG nova.network.neutron [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Refreshing network info cache for port a703ce6c-14a3-4a8e-aa09-8714098d603e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 916.876235] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:82:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a703ce6c-14a3-4a8e-aa09-8714098d603e', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 916.885480] env[61857]: DEBUG oslo.service.loopingcall [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.891578] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 916.892302] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9872dc6-6a6e-4706-b707-fa70876b19db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.920950] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 916.920950] env[61857]: value = "task-951240" [ 916.920950] env[61857]: _type = "Task" [ 916.920950] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.935269] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951240, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.080927] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951239, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.154673] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30912cd-08d3-4a98-8e83-08948b1595d3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.164169] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a84c4a-808e-435b-baa4-81c15acc39a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.207078] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011ae389-40a9-4f62-bb94-3b8e11a1b542 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.216821] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff00c0d5-e6fd-4ed8-be4d-55e3c46969c9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.233266] env[61857]: DEBUG nova.compute.provider_tree [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.348802] env[61857]: DEBUG nova.network.neutron [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Updated VIF entry in instance network info cache for port a703ce6c-14a3-4a8e-aa09-8714098d603e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 917.349235] env[61857]: DEBUG nova.network.neutron [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Updating instance_info_cache with network_info: [{"id": "a703ce6c-14a3-4a8e-aa09-8714098d603e", "address": "fa:16:3e:d1:82:65", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa703ce6c-14", "ovs_interfaceid": "a703ce6c-14a3-4a8e-aa09-8714098d603e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.431902] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951240, 'name': CreateVM_Task, 'duration_secs': 0.364012} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.432165] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 917.432927] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.433110] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.433548] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 917.433825] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5d07881-31b1-4ac6-b648-1681399b98db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.439842] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 917.439842] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b88367-3f3a-a7cb-13f4-c31ffa52626a" [ 917.439842] env[61857]: _type = "Task" [ 917.439842] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.448603] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b88367-3f3a-a7cb-13f4-c31ffa52626a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.575618] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951239, 'name': CloneVM_Task} progress is 95%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.737046] env[61857]: DEBUG nova.scheduler.client.report [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.856685] env[61857]: DEBUG oslo_concurrency.lockutils [req-c28b370b-cb41-46fe-8b77-ede1baa35255 req-ad66f5ad-d3ad-4ae6-bd8c-fc1b54027434 service nova] Releasing lock "refresh_cache-992bcab3-7396-495f-96a1-e99ecd13c961" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.951839] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b88367-3f3a-a7cb-13f4-c31ffa52626a, 'name': SearchDatastore_Task, 'duration_secs': 0.010031} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.952212] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.952454] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.952697] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.952852] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.953045] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.953318] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-115ea363-b14c-4e37-aa85-45181f296dd1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.964159] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.964159] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 917.964690] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52a2ba11-be29-46ae-95d2-a9c7a32fcc61 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.972668] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 917.972668] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ed28e9-1e25-8806-a9cb-be2da74d5ff1" [ 917.972668] env[61857]: _type = "Task" [ 917.972668] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.990022] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ed28e9-1e25-8806-a9cb-be2da74d5ff1, 'name': SearchDatastore_Task, 'duration_secs': 0.012269} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.990022] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0103405b-c2df-4cae-8c8d-41e5cb80ad1f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.996222] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 917.996222] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5254a879-1016-f2ed-846a-72b690239362" [ 917.996222] env[61857]: _type = "Task" [ 917.996222] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.008247] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5254a879-1016-f2ed-846a-72b690239362, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.073742] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951239, 'name': CloneVM_Task, 'duration_secs': 1.676892} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.074042] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Created linked-clone VM from snapshot [ 918.078019] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a4a21d-bded-4e8f-a867-1821e65ca9e8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.082839] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Uploading image efabb9a4-7907-4262-a138-86ad55c84b5b {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 918.112051] env[61857]: DEBUG oslo_vmware.rw_handles [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 918.112051] env[61857]: value = "vm-214159" [ 918.112051] env[61857]: _type = "VirtualMachine" [ 918.112051] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 918.112051] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-096d85fe-2084-404b-aad0-d17e9f2911d1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.119212] env[61857]: DEBUG oslo_vmware.rw_handles [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lease: (returnval){ [ 918.119212] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520398b7-0018-2568-1733-da5dc6a4d1dd" [ 918.119212] env[61857]: _type = "HttpNfcLease" [ 918.119212] env[61857]: } obtained for exporting VM: (result){ [ 918.119212] env[61857]: value = "vm-214159" [ 918.119212] env[61857]: _type = "VirtualMachine" [ 918.119212] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 918.119495] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the lease: (returnval){ [ 918.119495] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520398b7-0018-2568-1733-da5dc6a4d1dd" [ 918.119495] env[61857]: _type = "HttpNfcLease" [ 918.119495] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 918.126800] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 918.126800] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520398b7-0018-2568-1733-da5dc6a4d1dd" [ 918.126800] env[61857]: _type = "HttpNfcLease" [ 918.126800] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 918.243209] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.918s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.245922] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.027s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.248575] env[61857]: INFO nova.compute.claims [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.277014] env[61857]: INFO nova.scheduler.client.report [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleted allocations for instance 33cb5bbc-926d-42ee-b483-8d1e24707e40 [ 918.507902] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5254a879-1016-f2ed-846a-72b690239362, 'name': SearchDatastore_Task, 'duration_secs': 0.009942} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.508214] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.508493] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 992bcab3-7396-495f-96a1-e99ecd13c961/992bcab3-7396-495f-96a1-e99ecd13c961.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 918.508758] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-783d898b-c63e-4302-8807-cd6f84c9ebe0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.517586] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 918.517586] env[61857]: value = "task-951242" [ 918.517586] env[61857]: _type = "Task" [ 918.517586] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.526310] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951242, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.628616] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 918.628616] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520398b7-0018-2568-1733-da5dc6a4d1dd" [ 918.628616] env[61857]: _type = "HttpNfcLease" [ 918.628616] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 918.629042] env[61857]: DEBUG oslo_vmware.rw_handles [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 918.629042] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520398b7-0018-2568-1733-da5dc6a4d1dd" [ 918.629042] env[61857]: _type = "HttpNfcLease" [ 918.629042] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 918.629806] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-891d2e7e-e1b3-444a-805e-6c1284b7efb6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.638747] env[61857]: DEBUG oslo_vmware.rw_handles [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52663a47-eacd-ceed-0122-da0c3b98999e/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 918.638936] env[61857]: DEBUG oslo_vmware.rw_handles [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52663a47-eacd-ceed-0122-da0c3b98999e/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 918.742754] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-46957d17-0379-4f16-af20-63c79ce96d0a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.790169] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d527c8f6-ef53-4ab5-b1ae-398e6ec8e7f6 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "33cb5bbc-926d-42ee-b483-8d1e24707e40" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.376s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.034914] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951242, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.419894] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.420236] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.420496] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.420939] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.421142] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.423695] env[61857]: INFO nova.compute.manager [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Terminating instance [ 919.427009] env[61857]: DEBUG nova.compute.manager [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.427009] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 919.428243] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11f1bf8-8f2f-4310-917d-2f783f804138 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.437865] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 919.438361] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73707db0-f3a6-4271-a9a4-5b464e04568b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.453463] env[61857]: DEBUG oslo_vmware.api [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 919.453463] env[61857]: value = "task-951243" [ 919.453463] env[61857]: _type = "Task" [ 919.453463] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.473125] env[61857]: DEBUG oslo_vmware.api [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.536534] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951242, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.578564} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.542456] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 992bcab3-7396-495f-96a1-e99ecd13c961/992bcab3-7396-495f-96a1-e99ecd13c961.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 919.542456] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 919.542456] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-825377bc-974d-4c60-93fe-56addcd4b850 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.554687] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 919.554687] env[61857]: value = "task-951244" [ 919.554687] env[61857]: _type = "Task" [ 919.554687] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.569069] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951244, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.604962] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ece75e-46ef-4665-84eb-e44fae885c00 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.615916] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2261ae4-27e1-462e-aed8-985d67782b0e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.657874] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90986011-f0ec-434a-b7ae-8281964bd234 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.671188] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c72c52c-68e9-4948-bef4-fdae4c2543b4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.692379] env[61857]: DEBUG nova.compute.provider_tree [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.968789] env[61857]: DEBUG oslo_vmware.api [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951243, 'name': PowerOffVM_Task, 'duration_secs': 0.231477} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.968789] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 919.968789] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 919.968789] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa199a8b-0987-4efa-b075-c6844affba3e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.057541] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 920.057541] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 920.057541] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleting the datastore file [datastore1] e5e64477-a1aa-4a4d-91a2-b17d912e09c3 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 920.057541] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1fa7176-f185-4b4e-baad-63f8ab50a3ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.069172] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951244, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077578} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.070799] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.071265] env[61857]: DEBUG oslo_vmware.api [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 920.071265] env[61857]: value = "task-951246" [ 920.071265] env[61857]: _type = "Task" [ 920.071265] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.073870] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f42c11-fab5-419a-b300-66635d1b4d10 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.107268] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 992bcab3-7396-495f-96a1-e99ecd13c961/992bcab3-7396-495f-96a1-e99ecd13c961.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.112492] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aba8dbc3-7615-4915-a355-8ecbeefebb60 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.128744] env[61857]: DEBUG oslo_vmware.api [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951246, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.138549] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 920.138549] env[61857]: value = "task-951247" [ 920.138549] env[61857]: _type = "Task" [ 920.138549] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.147530] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951247, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.182904] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.182904] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.195345] env[61857]: DEBUG nova.scheduler.client.report [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.596041] env[61857]: DEBUG oslo_vmware.api [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951246, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196682} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.597208] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.597208] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 920.597208] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 920.597813] env[61857]: INFO nova.compute.manager [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Took 1.17 seconds to destroy the instance on the hypervisor. [ 920.600232] env[61857]: DEBUG oslo.service.loopingcall [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.600232] env[61857]: DEBUG nova.compute.manager [-] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 920.600232] env[61857]: DEBUG nova.network.neutron [-] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 920.654081] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951247, 'name': ReconfigVM_Task, 'duration_secs': 0.313219} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.659745] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 992bcab3-7396-495f-96a1-e99ecd13c961/992bcab3-7396-495f-96a1-e99ecd13c961.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.659745] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33be3e79-10ed-4fec-ab77-0188338eaaf7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.668161] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 920.668161] env[61857]: value = "task-951248" [ 920.668161] env[61857]: _type = "Task" [ 920.668161] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.679615] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951248, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.686027] env[61857]: DEBUG nova.compute.manager [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 920.706069] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.706626] env[61857]: DEBUG nova.compute.manager [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 920.709854] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.933s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.711170] env[61857]: DEBUG nova.objects.instance [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lazy-loading 'resources' on Instance uuid e88aed62-6561-4e29-b041-46b19cc5ef63 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.052358] env[61857]: DEBUG nova.compute.manager [req-9532a5b3-1e9a-4622-af70-72a08a2d5cb7 req-bb637b60-62d9-4068-b8de-0dbea7cca649 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Received event network-vif-deleted-0d52a5bd-50d8-41e1-86aa-b1eec00d8702 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 921.052578] env[61857]: INFO nova.compute.manager [req-9532a5b3-1e9a-4622-af70-72a08a2d5cb7 req-bb637b60-62d9-4068-b8de-0dbea7cca649 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Neutron deleted interface 0d52a5bd-50d8-41e1-86aa-b1eec00d8702; detaching it from the instance and deleting it from the info cache [ 921.052757] env[61857]: DEBUG nova.network.neutron [req-9532a5b3-1e9a-4622-af70-72a08a2d5cb7 req-bb637b60-62d9-4068-b8de-0dbea7cca649 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.136617] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquiring lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.136912] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.184105] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951248, 'name': Rename_Task, 'duration_secs': 0.163226} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.184400] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 921.184671] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e74df335-73ec-4837-8d99-620af9d7257b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.198679] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 921.198679] env[61857]: value = "task-951249" [ 921.198679] env[61857]: _type = "Task" [ 921.198679] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.212390] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951249, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.217965] env[61857]: DEBUG nova.compute.utils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.219865] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.220258] env[61857]: DEBUG nova.compute.manager [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 921.220450] env[61857]: DEBUG nova.network.neutron [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 921.274352] env[61857]: DEBUG nova.policy [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9056d344ff614950a60983450a2a40ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5653ed6ab0a948ef89bb03cf7426dd10', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 921.519620] env[61857]: DEBUG nova.network.neutron [-] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.531037] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.531037] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.556628] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-61ee557d-aac0-4d84-b316-7e131e3bfa68 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.568824] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8c1be9-cda3-43f0-bb24-2bd3163a23fe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.581359] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae721a8-ff97-44be-99d1-2d9ce80dd0a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.589456] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67be753-c5b5-4ef9-b6aa-56d21cf2078f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.605450] env[61857]: DEBUG nova.compute.manager [req-9532a5b3-1e9a-4622-af70-72a08a2d5cb7 req-bb637b60-62d9-4068-b8de-0dbea7cca649 service nova] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Detach interface failed, port_id=0d52a5bd-50d8-41e1-86aa-b1eec00d8702, reason: Instance e5e64477-a1aa-4a4d-91a2-b17d912e09c3 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 921.632246] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f5a89f-c0ae-48f9-8825-e24903e96087 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.640633] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748b3d5f-7cc0-48d0-8c29-168c69653b91 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.645236] env[61857]: DEBUG nova.compute.manager [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 921.658172] env[61857]: DEBUG nova.compute.provider_tree [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.710860] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951249, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.724442] env[61857]: DEBUG nova.compute.manager [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 921.770091] env[61857]: DEBUG nova.network.neutron [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Successfully created port: 9d3974fc-a398-4220-9f8c-d78aa99796bc {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.024647] env[61857]: INFO nova.compute.manager [-] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Took 1.43 seconds to deallocate network for instance. [ 922.034683] env[61857]: DEBUG nova.compute.utils [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.163046] env[61857]: DEBUG nova.scheduler.client.report [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.174430] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.212712] env[61857]: DEBUG oslo_vmware.api [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951249, 'name': PowerOnVM_Task, 'duration_secs': 0.818906} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.213022] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 922.213292] env[61857]: INFO nova.compute.manager [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Took 7.69 seconds to spawn the instance on the hypervisor. [ 922.213494] env[61857]: DEBUG nova.compute.manager [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.214759] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcfaa39-8240-4f99-96ff-e96c70600971 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.533811] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.539841] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.671285] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.962s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.673737] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.712s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.675296] env[61857]: INFO nova.compute.claims [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 922.692795] env[61857]: INFO nova.scheduler.client.report [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Deleted allocations for instance e88aed62-6561-4e29-b041-46b19cc5ef63 [ 922.718661] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.720174] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.733686] env[61857]: DEBUG nova.compute.manager [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 922.743849] env[61857]: INFO nova.compute.manager [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Took 29.90 seconds to build instance. [ 922.766599] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.766780] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.767276] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.767276] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.767470] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.767671] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.767929] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.768168] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.768412] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.768615] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.769259] env[61857]: DEBUG nova.virt.hardware [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.770097] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89e58bf-cddb-45b7-8bc7-a780b9cd457b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.780733] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4185b5-824c-4d1a-abfa-536de14f9c45 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.200477] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ebd9a549-42ea-44cd-add6-fdea4aea4b1f tempest-SecurityGroupsTestJSON-903839595 tempest-SecurityGroupsTestJSON-903839595-project-member] Lock "e88aed62-6561-4e29-b041-46b19cc5ef63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.738s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.224238] env[61857]: DEBUG nova.compute.manager [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.245394] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5f86116a-b172-4b71-83ec-d5181ff86ef7 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "992bcab3-7396-495f-96a1-e99ecd13c961" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.412s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.342742] env[61857]: DEBUG nova.compute.manager [req-5311e35a-370b-4a99-803c-079b86090031 req-0d4dc9c9-5f17-4651-aff5-760b5f97f066 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received event network-vif-plugged-9d3974fc-a398-4220-9f8c-d78aa99796bc {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 923.342742] env[61857]: DEBUG oslo_concurrency.lockutils [req-5311e35a-370b-4a99-803c-079b86090031 req-0d4dc9c9-5f17-4651-aff5-760b5f97f066 service nova] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.342742] env[61857]: DEBUG oslo_concurrency.lockutils [req-5311e35a-370b-4a99-803c-079b86090031 req-0d4dc9c9-5f17-4651-aff5-760b5f97f066 service nova] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.342742] env[61857]: DEBUG oslo_concurrency.lockutils [req-5311e35a-370b-4a99-803c-079b86090031 req-0d4dc9c9-5f17-4651-aff5-760b5f97f066 service nova] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.343550] env[61857]: DEBUG nova.compute.manager [req-5311e35a-370b-4a99-803c-079b86090031 req-0d4dc9c9-5f17-4651-aff5-760b5f97f066 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] No waiting events found dispatching network-vif-plugged-9d3974fc-a398-4220-9f8c-d78aa99796bc {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.343946] env[61857]: WARNING nova.compute.manager [req-5311e35a-370b-4a99-803c-079b86090031 req-0d4dc9c9-5f17-4651-aff5-760b5f97f066 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received unexpected event network-vif-plugged-9d3974fc-a398-4220-9f8c-d78aa99796bc for instance with vm_state building and task_state spawning. [ 923.495954] env[61857]: DEBUG nova.network.neutron [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Successfully updated port: 9d3974fc-a398-4220-9f8c-d78aa99796bc {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.605257] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.605585] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.605851] env[61857]: INFO nova.compute.manager [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Attaching volume 39d32790-20cc-4af1-a460-e9e10cb159b3 to /dev/sdb [ 923.651158] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd0ebba-6b2e-4b44-ae9d-4f265eb77784 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.659035] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1908d8c-217e-49b1-a11c-b29e34052e3f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.673050] env[61857]: DEBUG nova.virt.block_device [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updating existing volume attachment record: fde275be-5615-4b73-83ad-0f929dbacd50 {{(pid=61857) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 923.751873] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.999447] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.999612] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.999779] env[61857]: DEBUG nova.network.neutron [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 924.029845] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ccdc507-a968-4e32-9324-ed01eade0b30 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.042470] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783236d9-8f8d-4683-aa4e-d39af82f5be8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.077375] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581ddda6-63c2-4f39-8a68-dc4d88f147b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.085145] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be419cd-5d27-420b-935b-2d868f3fe77d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.100833] env[61857]: DEBUG nova.compute.provider_tree [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.200781] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "992bcab3-7396-495f-96a1-e99ecd13c961" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.201100] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "992bcab3-7396-495f-96a1-e99ecd13c961" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.201393] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "992bcab3-7396-495f-96a1-e99ecd13c961-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.201586] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "992bcab3-7396-495f-96a1-e99ecd13c961-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.201764] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "992bcab3-7396-495f-96a1-e99ecd13c961-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.204056] env[61857]: INFO nova.compute.manager [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Terminating instance [ 924.205881] env[61857]: DEBUG nova.compute.manager [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 924.206095] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 924.206927] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1eaeec4-2950-478e-a433-5a731e5c127d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.215071] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 924.215429] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b4a1709-8bbe-4e2b-97ce-c03f27a54845 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.222458] env[61857]: DEBUG oslo_vmware.api [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 924.222458] env[61857]: value = "task-951253" [ 924.222458] env[61857]: _type = "Task" [ 924.222458] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.230733] env[61857]: DEBUG oslo_vmware.api [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951253, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.549091] env[61857]: DEBUG nova.network.neutron [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 924.604484] env[61857]: DEBUG nova.scheduler.client.report [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.716303] env[61857]: DEBUG nova.network.neutron [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.733630] env[61857]: DEBUG oslo_vmware.api [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951253, 'name': PowerOffVM_Task, 'duration_secs': 0.216715} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.734042] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 924.735415] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 924.735415] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b041b16-57f7-4290-ac37-f9f130f3144f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.847916] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 924.848310] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 924.848576] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleting the datastore file [datastore2] 992bcab3-7396-495f-96a1-e99ecd13c961 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.848899] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dc87c7e0-d7f5-4035-a5e8-f4bf39ec56fa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.855555] env[61857]: DEBUG oslo_vmware.api [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 924.855555] env[61857]: value = "task-951255" [ 924.855555] env[61857]: _type = "Task" [ 924.855555] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.865969] env[61857]: DEBUG oslo_vmware.api [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.109559] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.110180] env[61857]: DEBUG nova.compute.manager [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 925.113657] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.316s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.114396] env[61857]: DEBUG nova.objects.instance [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lazy-loading 'resources' on Instance uuid 24d3d515-e03a-4b4e-bb8e-bc18537125ac {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.222021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.222021] env[61857]: DEBUG nova.compute.manager [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Instance network_info: |[{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 925.222021] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:00:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ffea58a-e043-4ad1-9ad7-45a61fdca001', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d3974fc-a398-4220-9f8c-d78aa99796bc', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 925.229257] env[61857]: DEBUG oslo.service.loopingcall [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.229700] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 925.230097] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e86cb49-bec1-46a9-a077-e4d5568bbc5e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.256817] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.256817] env[61857]: value = "task-951256" [ 925.256817] env[61857]: _type = "Task" [ 925.256817] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.267587] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951256, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.377839] env[61857]: DEBUG oslo_vmware.api [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.418151} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.378343] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.383112] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 925.383112] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 925.383112] env[61857]: INFO nova.compute.manager [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Took 1.18 seconds to destroy the instance on the hypervisor. [ 925.383342] env[61857]: DEBUG oslo.service.loopingcall [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.383588] env[61857]: DEBUG nova.compute.manager [-] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 925.383687] env[61857]: DEBUG nova.network.neutron [-] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 925.392156] env[61857]: DEBUG nova.compute.manager [req-cf3cc36b-2fc6-42b0-8b71-7609ea3b058c req-375649e8-bd45-4e6d-8d7c-9cdc5b0104f6 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received event network-changed-9d3974fc-a398-4220-9f8c-d78aa99796bc {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 925.392156] env[61857]: DEBUG nova.compute.manager [req-cf3cc36b-2fc6-42b0-8b71-7609ea3b058c req-375649e8-bd45-4e6d-8d7c-9cdc5b0104f6 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Refreshing instance network info cache due to event network-changed-9d3974fc-a398-4220-9f8c-d78aa99796bc. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 925.392525] env[61857]: DEBUG oslo_concurrency.lockutils [req-cf3cc36b-2fc6-42b0-8b71-7609ea3b058c req-375649e8-bd45-4e6d-8d7c-9cdc5b0104f6 service nova] Acquiring lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.392525] env[61857]: DEBUG oslo_concurrency.lockutils [req-cf3cc36b-2fc6-42b0-8b71-7609ea3b058c req-375649e8-bd45-4e6d-8d7c-9cdc5b0104f6 service nova] Acquired lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.392620] env[61857]: DEBUG nova.network.neutron [req-cf3cc36b-2fc6-42b0-8b71-7609ea3b058c req-375649e8-bd45-4e6d-8d7c-9cdc5b0104f6 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Refreshing network info cache for port 9d3974fc-a398-4220-9f8c-d78aa99796bc {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 925.615914] env[61857]: DEBUG nova.compute.utils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 925.618453] env[61857]: DEBUG nova.compute.manager [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 925.618826] env[61857]: DEBUG nova.network.neutron [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 925.690298] env[61857]: DEBUG nova.policy [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f719cbf1755d4c34bef92dab6679cb56', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ef111c8b3ff4f2383b4e81db026792b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 925.779516] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951256, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.976161] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4df24a-0b98-4a90-8260-ca256e2f1286 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.988152] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35291063-0638-42c8-9b85-b90a38cad276 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.020657] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debdf035-aa23-492d-84c0-1c7979bbc65e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.031528] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585a7b0f-02df-4eee-bfd1-b7b96aab9514 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.050356] env[61857]: DEBUG nova.compute.provider_tree [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.115160] env[61857]: DEBUG nova.network.neutron [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Successfully created port: 0b47d366-f4ff-4148-a2b4-be60b9228074 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.128049] env[61857]: DEBUG nova.compute.manager [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 926.150865] env[61857]: DEBUG nova.network.neutron [-] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.268988] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951256, 'name': CreateVM_Task, 'duration_secs': 0.690307} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.269176] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 926.273026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.273026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.273026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 926.273026] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c461ebf4-7779-4ff8-9389-a84d2b38557c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.275417] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 926.275417] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5265be91-0572-6e84-cdbb-a016f49a7eef" [ 926.275417] env[61857]: _type = "Task" [ 926.275417] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.284433] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5265be91-0572-6e84-cdbb-a016f49a7eef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.292224] env[61857]: DEBUG nova.network.neutron [req-cf3cc36b-2fc6-42b0-8b71-7609ea3b058c req-375649e8-bd45-4e6d-8d7c-9cdc5b0104f6 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updated VIF entry in instance network info cache for port 9d3974fc-a398-4220-9f8c-d78aa99796bc. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 926.292590] env[61857]: DEBUG nova.network.neutron [req-cf3cc36b-2fc6-42b0-8b71-7609ea3b058c req-375649e8-bd45-4e6d-8d7c-9cdc5b0104f6 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.557918] env[61857]: DEBUG nova.scheduler.client.report [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.651547] env[61857]: INFO nova.compute.manager [-] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Took 1.27 seconds to deallocate network for instance. [ 926.786280] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5265be91-0572-6e84-cdbb-a016f49a7eef, 'name': SearchDatastore_Task, 'duration_secs': 0.013795} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.786567] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.786858] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.787149] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.787311] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.787500] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.787778] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-907b1a29-b4ad-43b4-bab8-5532b50429dd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.796777] env[61857]: DEBUG oslo_concurrency.lockutils [req-cf3cc36b-2fc6-42b0-8b71-7609ea3b058c req-375649e8-bd45-4e6d-8d7c-9cdc5b0104f6 service nova] Releasing lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.797798] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.797990] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 926.798808] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c631822a-982b-40a4-a1af-997454127d0e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.804181] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 926.804181] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5271cbb8-c6c9-ec69-cd60-f4ef2610036c" [ 926.804181] env[61857]: _type = "Task" [ 926.804181] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.812220] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5271cbb8-c6c9-ec69-cd60-f4ef2610036c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.063290] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.950s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.066436] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.170s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.066436] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.066436] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 927.066877] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.677s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.066927] env[61857]: DEBUG nova.objects.instance [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lazy-loading 'resources' on Instance uuid 16b447b5-426a-4478-9d44-ae32b41dee50 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.069857] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6437deda-5bf8-4dc9-aed8-875269b7d587 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.078288] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c09a089-71e1-4501-bdf9-63ba8a893982 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.096137] env[61857]: INFO nova.scheduler.client.report [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Deleted allocations for instance 24d3d515-e03a-4b4e-bb8e-bc18537125ac [ 927.099659] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46c0e12-d752-4cf3-8e80-2fc329ff6069 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.110739] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d24156-d189-4a12-be44-a83bd75220f3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.146042] env[61857]: DEBUG nova.compute.manager [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 927.148519] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179428MB free_disk=40GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 927.148519] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.158073] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.191516] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.192451] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.192451] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.192451] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.192765] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.193067] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.193832] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.194490] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.194490] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.194623] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.194707] env[61857]: DEBUG nova.virt.hardware [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.199021] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58026879-e8a5-4b18-90f4-898da39fc04f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.205544] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38adf000-0c6b-4250-a96e-7108fc2be508 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.315482] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5271cbb8-c6c9-ec69-cd60-f4ef2610036c, 'name': SearchDatastore_Task, 'duration_secs': 0.010553} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.316568] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c93ccfe-fbf7-4b54-8833-7d04dc27742a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.325956] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 927.325956] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525f1353-8aab-3589-1042-e1bcce7743de" [ 927.325956] env[61857]: _type = "Task" [ 927.325956] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.332892] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525f1353-8aab-3589-1042-e1bcce7743de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.417223] env[61857]: DEBUG nova.compute.manager [req-22b2bc19-1825-4291-a096-090e0a763b7a req-3c6d828a-800f-4bfe-bd6f-114cdf4e1e2e service nova] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Received event network-vif-deleted-a703ce6c-14a3-4a8e-aa09-8714098d603e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 927.608547] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7e8cc55-93b5-4e6c-8bdf-b709536cd7c6 tempest-FloatingIPsAssociationTestJSON-401439786 tempest-FloatingIPsAssociationTestJSON-401439786-project-member] Lock "24d3d515-e03a-4b4e-bb8e-bc18537125ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.992s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.841198] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525f1353-8aab-3589-1042-e1bcce7743de, 'name': SearchDatastore_Task, 'duration_secs': 0.018514} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.841476] env[61857]: DEBUG nova.network.neutron [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Successfully updated port: 0b47d366-f4ff-4148-a2b4-be60b9228074 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 927.844927] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.845865] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] f982f0c5-c701-4e5b-b96d-c21a3c0da24f/f982f0c5-c701-4e5b-b96d-c21a3c0da24f.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 927.848349] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-42ad36a6-4880-4e30-8780-51239ff0ea7c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.856240] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 927.856240] env[61857]: value = "task-951258" [ 927.856240] env[61857]: _type = "Task" [ 927.856240] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.857458] env[61857]: DEBUG oslo_vmware.rw_handles [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52663a47-eacd-ceed-0122-da0c3b98999e/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 927.858453] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c9fe032-5fba-446c-ae04-eb6d64ca47ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.881075] env[61857]: DEBUG oslo_vmware.rw_handles [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52663a47-eacd-ceed-0122-da0c3b98999e/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 927.881294] env[61857]: ERROR oslo_vmware.rw_handles [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52663a47-eacd-ceed-0122-da0c3b98999e/disk-0.vmdk due to incomplete transfer. [ 927.881585] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951258, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.888599] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f32d7298-9248-43ec-903a-ef960263f424 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.897259] env[61857]: DEBUG oslo_vmware.rw_handles [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52663a47-eacd-ceed-0122-da0c3b98999e/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 927.897547] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Uploaded image efabb9a4-7907-4262-a138-86ad55c84b5b to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 927.899961] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 927.900581] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f445df94-6f5f-4b1b-9632-9ab096aeb588 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.906329] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 927.906329] env[61857]: value = "task-951259" [ 927.906329] env[61857]: _type = "Task" [ 927.906329] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.912007] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5d2991-b99a-49ae-a662-8f05db1ea6db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.920031] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951259, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.922514] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d36314-dff9-4c5e-9420-0c08939af042 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.954191] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf644cf5-140e-465c-be4f-4c40e8a4c7a2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.962164] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6fb031-b42e-4853-adb5-6cce3e0c80c4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.975397] env[61857]: DEBUG nova.compute.provider_tree [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.224373] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Volume attach. Driver type: vmdk {{(pid=61857) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 928.224694] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214162', 'volume_id': '39d32790-20cc-4af1-a460-e9e10cb159b3', 'name': 'volume-39d32790-20cc-4af1-a460-e9e10cb159b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd9d05629-79ef-45c6-ac54-b3bc18d306f4', 'attached_at': '', 'detached_at': '', 'volume_id': '39d32790-20cc-4af1-a460-e9e10cb159b3', 'serial': '39d32790-20cc-4af1-a460-e9e10cb159b3'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 928.225762] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0571f81-1e4e-41b2-a0c9-2f90f593f289 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.244139] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0eb1f79-aea6-41e4-b359-063fc9e59e94 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.273446] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] volume-39d32790-20cc-4af1-a460-e9e10cb159b3/volume-39d32790-20cc-4af1-a460-e9e10cb159b3.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.273845] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b3389bd-a524-4907-a5ea-17ff7f18c5d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.292376] env[61857]: DEBUG oslo_vmware.api [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 928.292376] env[61857]: value = "task-951260" [ 928.292376] env[61857]: _type = "Task" [ 928.292376] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.301256] env[61857]: DEBUG oslo_vmware.api [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951260, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.350628] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-d85f8a2d-6053-45ec-81bb-de9fe02af3fe" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.350628] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-d85f8a2d-6053-45ec-81bb-de9fe02af3fe" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.350628] env[61857]: DEBUG nova.network.neutron [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 928.372337] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951258, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.415747] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951259, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.478104] env[61857]: DEBUG nova.scheduler.client.report [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.801817] env[61857]: DEBUG oslo_vmware.api [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951260, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.869877] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951258, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.615521} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.870183] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] f982f0c5-c701-4e5b-b96d-c21a3c0da24f/f982f0c5-c701-4e5b-b96d-c21a3c0da24f.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 928.870564] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.870840] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f8af31e-6815-4e90-9a38-3f96bc00cbac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.877545] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 928.877545] env[61857]: value = "task-951261" [ 928.877545] env[61857]: _type = "Task" [ 928.877545] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.885021] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951261, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.897521] env[61857]: DEBUG nova.network.neutron [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 928.916382] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951259, 'name': Destroy_Task, 'duration_secs': 0.694022} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.916649] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Destroyed the VM [ 928.916886] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 928.917155] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9ee5803e-c227-40cf-b96c-286162c6d8a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.923766] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 928.923766] env[61857]: value = "task-951262" [ 928.923766] env[61857]: _type = "Task" [ 928.923766] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.931614] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951262, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.983860] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.917s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.986934] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.491s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.988120] env[61857]: INFO nova.compute.claims [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.012634] env[61857]: INFO nova.scheduler.client.report [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Deleted allocations for instance 16b447b5-426a-4478-9d44-ae32b41dee50 [ 929.096861] env[61857]: DEBUG nova.network.neutron [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Updating instance_info_cache with network_info: [{"id": "0b47d366-f4ff-4148-a2b4-be60b9228074", "address": "fa:16:3e:82:6f:b3", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b47d366-f4", "ovs_interfaceid": "0b47d366-f4ff-4148-a2b4-be60b9228074", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.305719] env[61857]: DEBUG oslo_vmware.api [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951260, 'name': ReconfigVM_Task, 'duration_secs': 0.511667} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.305719] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Reconfigured VM instance instance-0000004d to attach disk [datastore2] volume-39d32790-20cc-4af1-a460-e9e10cb159b3/volume-39d32790-20cc-4af1-a460-e9e10cb159b3.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.308225] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a02099c9-6b75-439d-9cb8-fae1fb4c84fb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.325107] env[61857]: DEBUG oslo_vmware.api [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 929.325107] env[61857]: value = "task-951263" [ 929.325107] env[61857]: _type = "Task" [ 929.325107] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.335554] env[61857]: DEBUG oslo_vmware.api [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951263, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.386891] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951261, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.129364} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.389026] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 929.389026] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abd2986-122b-43f6-be2f-e5f3a74e25ff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.414062] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] f982f0c5-c701-4e5b-b96d-c21a3c0da24f/f982f0c5-c701-4e5b-b96d-c21a3c0da24f.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 929.415285] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73f52283-aea8-49c3-b1de-96b7a0211ebb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.437484] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951262, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.439146] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 929.439146] env[61857]: value = "task-951264" [ 929.439146] env[61857]: _type = "Task" [ 929.439146] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.447930] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951264, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.455903] env[61857]: DEBUG nova.compute.manager [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Received event network-vif-plugged-0b47d366-f4ff-4148-a2b4-be60b9228074 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 929.456145] env[61857]: DEBUG oslo_concurrency.lockutils [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] Acquiring lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.456412] env[61857]: DEBUG oslo_concurrency.lockutils [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] Lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.456535] env[61857]: DEBUG oslo_concurrency.lockutils [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] Lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.456687] env[61857]: DEBUG nova.compute.manager [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] No waiting events found dispatching network-vif-plugged-0b47d366-f4ff-4148-a2b4-be60b9228074 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.456855] env[61857]: WARNING nova.compute.manager [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Received unexpected event network-vif-plugged-0b47d366-f4ff-4148-a2b4-be60b9228074 for instance with vm_state building and task_state spawning. [ 929.457034] env[61857]: DEBUG nova.compute.manager [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Received event network-changed-0b47d366-f4ff-4148-a2b4-be60b9228074 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 929.457186] env[61857]: DEBUG nova.compute.manager [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Refreshing instance network info cache due to event network-changed-0b47d366-f4ff-4148-a2b4-be60b9228074. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 929.457354] env[61857]: DEBUG oslo_concurrency.lockutils [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] Acquiring lock "refresh_cache-d85f8a2d-6053-45ec-81bb-de9fe02af3fe" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.524289] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eeca531d-e859-4b0e-8884-05caee495820 tempest-AttachVolumeTestJSON-1922554849 tempest-AttachVolumeTestJSON-1922554849-project-member] Lock "16b447b5-426a-4478-9d44-ae32b41dee50" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.715s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.600103] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-d85f8a2d-6053-45ec-81bb-de9fe02af3fe" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.600103] env[61857]: DEBUG nova.compute.manager [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Instance network_info: |[{"id": "0b47d366-f4ff-4148-a2b4-be60b9228074", "address": "fa:16:3e:82:6f:b3", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b47d366-f4", "ovs_interfaceid": "0b47d366-f4ff-4148-a2b4-be60b9228074", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 929.600350] env[61857]: DEBUG oslo_concurrency.lockutils [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] Acquired lock "refresh_cache-d85f8a2d-6053-45ec-81bb-de9fe02af3fe" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.600443] env[61857]: DEBUG nova.network.neutron [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Refreshing network info cache for port 0b47d366-f4ff-4148-a2b4-be60b9228074 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 929.602669] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:6f:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b47d366-f4ff-4148-a2b4-be60b9228074', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 929.615945] env[61857]: DEBUG oslo.service.loopingcall [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.619201] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 929.619782] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d92d6548-7de7-417d-8004-c7e9a9d19c2d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.656055] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 929.656055] env[61857]: value = "task-951265" [ 929.656055] env[61857]: _type = "Task" [ 929.656055] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.667379] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951265, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.836389] env[61857]: DEBUG oslo_vmware.api [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951263, 'name': ReconfigVM_Task, 'duration_secs': 0.160857} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.836759] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214162', 'volume_id': '39d32790-20cc-4af1-a460-e9e10cb159b3', 'name': 'volume-39d32790-20cc-4af1-a460-e9e10cb159b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd9d05629-79ef-45c6-ac54-b3bc18d306f4', 'attached_at': '', 'detached_at': '', 'volume_id': '39d32790-20cc-4af1-a460-e9e10cb159b3', 'serial': '39d32790-20cc-4af1-a460-e9e10cb159b3'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 929.903214] env[61857]: DEBUG nova.network.neutron [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Updated VIF entry in instance network info cache for port 0b47d366-f4ff-4148-a2b4-be60b9228074. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 929.903657] env[61857]: DEBUG nova.network.neutron [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Updating instance_info_cache with network_info: [{"id": "0b47d366-f4ff-4148-a2b4-be60b9228074", "address": "fa:16:3e:82:6f:b3", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b47d366-f4", "ovs_interfaceid": "0b47d366-f4ff-4148-a2b4-be60b9228074", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.938160] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951262, 'name': RemoveSnapshot_Task, 'duration_secs': 0.656258} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.938462] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 929.938735] env[61857]: DEBUG nova.compute.manager [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.939610] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea23851-f8ef-4de3-b1af-2ae166ec8321 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.955183] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951264, 'name': ReconfigVM_Task, 'duration_secs': 0.28885} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.955183] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Reconfigured VM instance instance-00000054 to attach disk [datastore2] f982f0c5-c701-4e5b-b96d-c21a3c0da24f/f982f0c5-c701-4e5b-b96d-c21a3c0da24f.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.955446] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01a94ed8-fd18-4e13-9408-37c6693c5492 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.960973] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 929.960973] env[61857]: value = "task-951267" [ 929.960973] env[61857]: _type = "Task" [ 929.960973] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.969317] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951267, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.165652] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951265, 'name': CreateVM_Task, 'duration_secs': 0.344137} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.165848] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 930.166515] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.166679] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.166998] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.169536] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9709271-1875-41ed-bf1e-505f409187ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.175047] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 930.175047] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e9d26b-4ae5-9ae3-3e46-3537253d6aa3" [ 930.175047] env[61857]: _type = "Task" [ 930.175047] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.182993] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e9d26b-4ae5-9ae3-3e46-3537253d6aa3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.309792] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89f0a28-08c8-4a57-97d8-bbafadaebd83 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.317025] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494d10b2-faa3-4973-939b-7edeb2b2fb7b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.351027] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311b1407-5193-43a7-8bb9-25ac381365ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.358510] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b55956b-8a60-48bd-b045-afbb7187305f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.372018] env[61857]: DEBUG nova.compute.provider_tree [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.407115] env[61857]: DEBUG oslo_concurrency.lockutils [req-98b1bac0-785d-4283-a242-0e6faa7f0299 req-5fd2598c-ecf4-4779-9ecc-80a52f4406f8 service nova] Releasing lock "refresh_cache-d85f8a2d-6053-45ec-81bb-de9fe02af3fe" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.457680] env[61857]: INFO nova.compute.manager [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Shelve offloading [ 930.461438] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 930.462062] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b3aadf6-b7d9-4081-97c1-69fc62d00675 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.472769] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951267, 'name': Rename_Task, 'duration_secs': 0.143309} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.473974] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 930.474310] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 930.474310] env[61857]: value = "task-951268" [ 930.474310] env[61857]: _type = "Task" [ 930.474310] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.474496] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-278911fb-6887-4d51-9df5-6fd638fa261a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.485128] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] VM already powered off {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 930.485803] env[61857]: DEBUG nova.compute.manager [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.485803] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 930.485803] env[61857]: value = "task-951269" [ 930.485803] env[61857]: _type = "Task" [ 930.485803] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.486293] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d1eaf8-78b5-4b29-851b-ce1f827bf38e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.498151] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.498272] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.498439] env[61857]: DEBUG nova.network.neutron [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 930.499800] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951269, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.687453] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e9d26b-4ae5-9ae3-3e46-3537253d6aa3, 'name': SearchDatastore_Task, 'duration_secs': 0.009313} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.687938] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.688197] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 930.689111] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.689111] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.689111] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 930.689475] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4197a997-c8dc-40ed-85ac-93c694e996a0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.698661] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 930.698960] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 930.699848] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b012a839-91cb-4644-b157-fe18f493a557 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.706643] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 930.706643] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521da216-b3a4-1b88-e411-8fcf646d47f3" [ 930.706643] env[61857]: _type = "Task" [ 930.706643] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.713747] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521da216-b3a4-1b88-e411-8fcf646d47f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.877774] env[61857]: DEBUG nova.scheduler.client.report [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.889209] env[61857]: DEBUG nova.objects.instance [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lazy-loading 'flavor' on Instance uuid d9d05629-79ef-45c6-ac54-b3bc18d306f4 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.003272] env[61857]: DEBUG oslo_vmware.api [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951269, 'name': PowerOnVM_Task, 'duration_secs': 0.486197} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.003272] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 931.003272] env[61857]: INFO nova.compute.manager [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Took 8.27 seconds to spawn the instance on the hypervisor. [ 931.003272] env[61857]: DEBUG nova.compute.manager [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.004090] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-935cfd08-122f-4471-a51d-924da7219cbe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.084728] env[61857]: INFO nova.compute.manager [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Rescuing [ 931.085159] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.085803] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.086077] env[61857]: DEBUG nova.network.neutron [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 931.216024] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521da216-b3a4-1b88-e411-8fcf646d47f3, 'name': SearchDatastore_Task, 'duration_secs': 0.008969} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.216798] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b84ae4e2-0f5c-47a4-9fc4-97aebae26e8a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.221744] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 931.221744] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52649083-93b8-d319-517e-b4bfa221c7ba" [ 931.221744] env[61857]: _type = "Task" [ 931.221744] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.228919] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52649083-93b8-d319-517e-b4bfa221c7ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.263674] env[61857]: DEBUG nova.network.neutron [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updating instance_info_cache with network_info: [{"id": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "address": "fa:16:3e:dc:03:c9", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5da8191c-5f", "ovs_interfaceid": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.383425] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.383965] env[61857]: DEBUG nova.compute.manager [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 931.386766] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.473s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.387073] env[61857]: DEBUG nova.objects.instance [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lazy-loading 'resources' on Instance uuid d979a8a1-a5b5-428e-9b60-7d827337218f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.394485] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae46309e-2fe7-4e1f-a4a7-049d85a5f7d9 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.788s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.522884] env[61857]: INFO nova.compute.manager [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Took 27.33 seconds to build instance. [ 931.732454] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52649083-93b8-d319-517e-b4bfa221c7ba, 'name': SearchDatastore_Task, 'duration_secs': 0.009108} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.732454] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.732454] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] d85f8a2d-6053-45ec-81bb-de9fe02af3fe/d85f8a2d-6053-45ec-81bb-de9fe02af3fe.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 931.732761] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30e73b38-e50f-4111-969d-e3f8d29de8f6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.739055] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 931.739055] env[61857]: value = "task-951270" [ 931.739055] env[61857]: _type = "Task" [ 931.739055] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.749197] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.767630] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.893025] env[61857]: DEBUG nova.compute.utils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.895844] env[61857]: DEBUG nova.compute.manager [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 931.896086] env[61857]: DEBUG nova.network.neutron [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 931.995396] env[61857]: DEBUG nova.network.neutron [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updating instance_info_cache with network_info: [{"id": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "address": "fa:16:3e:94:19:6b", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9f80f94-9b", "ovs_interfaceid": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.029440] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6ed59548-5169-42ec-a617-adef4d274af5 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.844s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.039385] env[61857]: DEBUG nova.policy [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e428c8cc5124262a2f51f71080e104c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '36863339760945dc89398dad43684440', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.230264] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b376d4b7-fb31-4722-b6fc-36b15d73468b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.239265] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1848d3f3-9cd3-450d-9336-613c61ea7dc8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.259888] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951270, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.293069] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e183dbf8-9832-4a0f-8cdf-037c3fc42a5d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.304630] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ab1e89-75d2-48cf-b60e-61c1eb6e99dd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.317860] env[61857]: DEBUG nova.compute.provider_tree [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.396320] env[61857]: DEBUG nova.compute.manager [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 932.433104] env[61857]: DEBUG nova.compute.manager [req-0b765de9-6af9-4c52-9e4a-5b2fba6d5210 req-1d8191e3-f04f-4d5f-a945-eb693e434e48 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Received event network-vif-unplugged-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.433523] env[61857]: DEBUG oslo_concurrency.lockutils [req-0b765de9-6af9-4c52-9e4a-5b2fba6d5210 req-1d8191e3-f04f-4d5f-a945-eb693e434e48 service nova] Acquiring lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.437407] env[61857]: DEBUG oslo_concurrency.lockutils [req-0b765de9-6af9-4c52-9e4a-5b2fba6d5210 req-1d8191e3-f04f-4d5f-a945-eb693e434e48 service nova] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.437407] env[61857]: DEBUG oslo_concurrency.lockutils [req-0b765de9-6af9-4c52-9e4a-5b2fba6d5210 req-1d8191e3-f04f-4d5f-a945-eb693e434e48 service nova] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.437407] env[61857]: DEBUG nova.compute.manager [req-0b765de9-6af9-4c52-9e4a-5b2fba6d5210 req-1d8191e3-f04f-4d5f-a945-eb693e434e48 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] No waiting events found dispatching network-vif-unplugged-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.437407] env[61857]: WARNING nova.compute.manager [req-0b765de9-6af9-4c52-9e4a-5b2fba6d5210 req-1d8191e3-f04f-4d5f-a945-eb693e434e48 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Received unexpected event network-vif-unplugged-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 for instance with vm_state shelved and task_state shelving_offloading. [ 932.499736] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.579593] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 932.580585] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd42158-ae51-4924-bcd4-c0860ee7b17d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.592353] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 932.592626] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-967cdcb2-9dc1-4a5d-9b9b-8c356f9477c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.734982] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 932.735218] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 932.735362] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleting the datastore file [datastore2] 781a2790-e317-46fe-9be8-40c9e5f1f771 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 932.735627] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ac7211b-3402-4d00-826e-b4359aa64e9e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.741717] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 932.741717] env[61857]: value = "task-951273" [ 932.741717] env[61857]: _type = "Task" [ 932.741717] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.749535] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951273, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.757583] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951270, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553228} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.758471] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] d85f8a2d-6053-45ec-81bb-de9fe02af3fe/d85f8a2d-6053-45ec-81bb-de9fe02af3fe.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 932.759047] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.759047] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6aa1530d-230e-49bb-bdea-060675c31896 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.765279] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 932.765279] env[61857]: value = "task-951274" [ 932.765279] env[61857]: _type = "Task" [ 932.765279] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.773911] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951274, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.815181] env[61857]: DEBUG nova.network.neutron [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Successfully created port: 553edf3c-e40b-4c94-b91c-f81a54df6643 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.823493] env[61857]: DEBUG nova.scheduler.client.report [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.031411] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 933.032165] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b86901f-f251-45d1-a364-d9740755665a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.040810] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 933.040810] env[61857]: value = "task-951275" [ 933.040810] env[61857]: _type = "Task" [ 933.040810] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.050801] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951275, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.257624] env[61857]: DEBUG oslo_vmware.api [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951273, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14007} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.257926] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.258134] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 933.258319] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 933.275728] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951274, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075198} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.275728] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.276458] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7ae3a8-471c-463a-b307-df8a0ff62f4e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.304938] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] d85f8a2d-6053-45ec-81bb-de9fe02af3fe/d85f8a2d-6053-45ec-81bb-de9fe02af3fe.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.309049] env[61857]: INFO nova.scheduler.client.report [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleted allocations for instance 781a2790-e317-46fe-9be8-40c9e5f1f771 [ 933.309049] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-086434ca-adc4-45c1-9044-ab19388428c8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.329777] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.941s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.334505] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 18.537s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.340026] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 933.340026] env[61857]: value = "task-951276" [ 933.340026] env[61857]: _type = "Task" [ 933.340026] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.349780] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951276, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.365037] env[61857]: INFO nova.scheduler.client.report [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted allocations for instance d979a8a1-a5b5-428e-9b60-7d827337218f [ 933.415750] env[61857]: DEBUG nova.compute.manager [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 933.458941] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.459458] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.459458] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.459600] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.461471] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.461471] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.461471] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.461471] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.461471] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.461471] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.461471] env[61857]: DEBUG nova.virt.hardware [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.462752] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e29ab61-1f75-47cd-b2f9-54ce6d77b128 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.473088] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c452a8a-aa71-4bfe-a0e0-22b1b0ea7319 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.551393] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951275, 'name': PowerOffVM_Task, 'duration_secs': 0.206491} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.551869] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 933.552675] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98e02c5-fcbe-431a-a089-99a641efc6c3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.576401] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c700636-4fe7-4955-97ab-3e3fa7707a05 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.607833] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 933.608139] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6eed2ab1-b0d5-454f-97bf-502ac5180a4b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.614613] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 933.614613] env[61857]: value = "task-951277" [ 933.614613] env[61857]: _type = "Task" [ 933.614613] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.623120] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.826055] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.841871] env[61857]: INFO nova.compute.claims [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.853505] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951276, 'name': ReconfigVM_Task, 'duration_secs': 0.367803} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.854146] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Reconfigured VM instance instance-00000055 to attach disk [datastore2] d85f8a2d-6053-45ec-81bb-de9fe02af3fe/d85f8a2d-6053-45ec-81bb-de9fe02af3fe.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.854390] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e58f0f61-11b4-40d6-af8f-a333aee94ca5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.861323] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 933.861323] env[61857]: value = "task-951278" [ 933.861323] env[61857]: _type = "Task" [ 933.861323] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.869683] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951278, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.878669] env[61857]: DEBUG oslo_concurrency.lockutils [None req-73b266fd-b3eb-466f-bab2-2e967bb51128 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "d979a8a1-a5b5-428e-9b60-7d827337218f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.888s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.125950] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] VM already powered off {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 934.126271] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.126438] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.126618] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.126941] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.127370] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6658f153-fb0a-4a47-9bb8-547f738aadab {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.137254] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.137454] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 934.138336] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b96a6e45-332c-4d8c-8381-011136698652 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.144552] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 934.144552] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ec0658-6142-82fe-6aec-a887a51bf506" [ 934.144552] env[61857]: _type = "Task" [ 934.144552] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.152756] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ec0658-6142-82fe-6aec-a887a51bf506, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.351036] env[61857]: INFO nova.compute.resource_tracker [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating resource usage from migration c72d05c4-0844-486a-bf03-11222909c20c [ 934.370836] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951278, 'name': Rename_Task, 'duration_secs': 0.244485} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.373808] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 934.376553] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78654e15-194e-43e6-9b86-52f3126c9146 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.388713] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 934.388713] env[61857]: value = "task-951279" [ 934.388713] env[61857]: _type = "Task" [ 934.388713] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.399259] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951279, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.478779] env[61857]: DEBUG nova.compute.manager [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Received event network-changed-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 934.478995] env[61857]: DEBUG nova.compute.manager [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Refreshing instance network info cache due to event network-changed-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 934.480404] env[61857]: DEBUG oslo_concurrency.lockutils [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] Acquiring lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.480404] env[61857]: DEBUG oslo_concurrency.lockutils [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] Acquired lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.480404] env[61857]: DEBUG nova.network.neutron [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Refreshing network info cache for port 5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 934.617632] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5876aab8-85f2-4fdd-a62f-36d69c09d07c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.626322] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322e0cbb-e3e9-4e46-a37e-52ed94c09970 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.661765] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905dac9b-2478-41c3-99d6-dfbe7ff34fbc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.673549] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854faa6e-bb6c-4689-aea9-37cea6832316 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.677840] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ec0658-6142-82fe-6aec-a887a51bf506, 'name': SearchDatastore_Task, 'duration_secs': 0.02433} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.679253] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adb8d2f1-5f85-4e71-816e-9254de2b77f9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.690315] env[61857]: DEBUG nova.compute.provider_tree [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.695127] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 934.695127] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5267f139-525e-5920-ea9e-9685cf0806a7" [ 934.695127] env[61857]: _type = "Task" [ 934.695127] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.703161] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5267f139-525e-5920-ea9e-9685cf0806a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.865091] env[61857]: DEBUG nova.network.neutron [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Successfully updated port: 553edf3c-e40b-4c94-b91c-f81a54df6643 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.899596] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951279, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.196339] env[61857]: DEBUG nova.scheduler.client.report [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.213027] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5267f139-525e-5920-ea9e-9685cf0806a7, 'name': SearchDatastore_Task, 'duration_secs': 0.012558} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.214613] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.215201] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] d9d05629-79ef-45c6-ac54-b3bc18d306f4/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk. {{(pid=61857) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 935.215847] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76ec2d1a-ae2a-47e4-899b-54ca893a34ab {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.223930] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 935.223930] env[61857]: value = "task-951280" [ 935.223930] env[61857]: _type = "Task" [ 935.223930] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.235769] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951280, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.241029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "781a2790-e317-46fe-9be8-40c9e5f1f771" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.337018] env[61857]: DEBUG nova.network.neutron [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updated VIF entry in instance network info cache for port 5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 935.337400] env[61857]: DEBUG nova.network.neutron [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updating instance_info_cache with network_info: [{"id": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "address": "fa:16:3e:dc:03:c9", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap5da8191c-5f", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.367159] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquiring lock "refresh_cache-825f5d87-76b1-4e4e-9b3d-4be419c7d323" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.367258] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquired lock "refresh_cache-825f5d87-76b1-4e4e-9b3d-4be419c7d323" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.368131] env[61857]: DEBUG nova.network.neutron [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 935.402449] env[61857]: DEBUG oslo_vmware.api [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951279, 'name': PowerOnVM_Task, 'duration_secs': 0.773588} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.402973] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 935.403336] env[61857]: INFO nova.compute.manager [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Took 8.26 seconds to spawn the instance on the hypervisor. [ 935.404608] env[61857]: DEBUG nova.compute.manager [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.406294] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33a22df-305d-43c3-a7b6-26292ce6c674 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.709026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.374s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.709026] env[61857]: INFO nova.compute.manager [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Migrating [ 935.714564] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.495s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.716407] env[61857]: INFO nova.compute.claims [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.738798] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951280, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444281} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.738798] env[61857]: INFO nova.virt.vmwareapi.ds_util [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] d9d05629-79ef-45c6-ac54-b3bc18d306f4/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk. [ 935.738798] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36269f14-a0af-4645-acda-7e43b9e18021 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.767924] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] d9d05629-79ef-45c6-ac54-b3bc18d306f4/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.769050] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27ae0ba8-820a-4f04-95c9-3401691c8aa2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.793054] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 935.793054] env[61857]: value = "task-951281" [ 935.793054] env[61857]: _type = "Task" [ 935.793054] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.799727] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951281, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.839771] env[61857]: DEBUG oslo_concurrency.lockutils [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] Releasing lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.840077] env[61857]: DEBUG nova.compute.manager [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received event network-changed-9d3974fc-a398-4220-9f8c-d78aa99796bc {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.840384] env[61857]: DEBUG nova.compute.manager [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Refreshing instance network info cache due to event network-changed-9d3974fc-a398-4220-9f8c-d78aa99796bc. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 935.840548] env[61857]: DEBUG oslo_concurrency.lockutils [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] Acquiring lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.840697] env[61857]: DEBUG oslo_concurrency.lockutils [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] Acquired lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.840951] env[61857]: DEBUG nova.network.neutron [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Refreshing network info cache for port 9d3974fc-a398-4220-9f8c-d78aa99796bc {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.928874] env[61857]: INFO nova.compute.manager [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Took 25.99 seconds to build instance. [ 935.941602] env[61857]: DEBUG nova.network.neutron [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 936.232025] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.232025] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.232025] env[61857]: DEBUG nova.network.neutron [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 936.301744] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951281, 'name': ReconfigVM_Task, 'duration_secs': 0.348322} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.302393] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Reconfigured VM instance instance-0000004d to attach disk [datastore1] d9d05629-79ef-45c6-ac54-b3bc18d306f4/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.305182] env[61857]: DEBUG nova.network.neutron [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Updating instance_info_cache with network_info: [{"id": "553edf3c-e40b-4c94-b91c-f81a54df6643", "address": "fa:16:3e:9b:83:91", "network": {"id": "60998eff-d613-4674-b164-c141b817e4a0", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-175154447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36863339760945dc89398dad43684440", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap553edf3c-e4", "ovs_interfaceid": "553edf3c-e40b-4c94-b91c-f81a54df6643", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.312162] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9218404-08cb-4cb8-9919-a9e3f351136b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.347290] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3152c28e-d956-4a5e-84d8-b11858ba34e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.366113] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 936.366113] env[61857]: value = "task-951282" [ 936.366113] env[61857]: _type = "Task" [ 936.366113] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.376782] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951282, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.397290] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "c3c9df36-d523-4a0f-9484-4446ad8c4138" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.397546] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.431089] env[61857]: DEBUG oslo_concurrency.lockutils [None req-06b552e5-a9a7-49ec-aa38-85304862b447 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.501s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.559014] env[61857]: DEBUG nova.compute.manager [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Received event network-vif-plugged-553edf3c-e40b-4c94-b91c-f81a54df6643 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 936.559945] env[61857]: DEBUG oslo_concurrency.lockutils [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] Acquiring lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.560263] env[61857]: DEBUG oslo_concurrency.lockutils [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] Lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.560532] env[61857]: DEBUG oslo_concurrency.lockutils [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] Lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.560750] env[61857]: DEBUG nova.compute.manager [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] No waiting events found dispatching network-vif-plugged-553edf3c-e40b-4c94-b91c-f81a54df6643 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 936.560991] env[61857]: WARNING nova.compute.manager [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Received unexpected event network-vif-plugged-553edf3c-e40b-4c94-b91c-f81a54df6643 for instance with vm_state building and task_state spawning. [ 936.561682] env[61857]: DEBUG nova.compute.manager [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Received event network-changed-553edf3c-e40b-4c94-b91c-f81a54df6643 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 936.561917] env[61857]: DEBUG nova.compute.manager [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Refreshing instance network info cache due to event network-changed-553edf3c-e40b-4c94-b91c-f81a54df6643. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 936.562155] env[61857]: DEBUG oslo_concurrency.lockutils [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] Acquiring lock "refresh_cache-825f5d87-76b1-4e4e-9b3d-4be419c7d323" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.814102] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Releasing lock "refresh_cache-825f5d87-76b1-4e4e-9b3d-4be419c7d323" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.814422] env[61857]: DEBUG nova.compute.manager [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Instance network_info: |[{"id": "553edf3c-e40b-4c94-b91c-f81a54df6643", "address": "fa:16:3e:9b:83:91", "network": {"id": "60998eff-d613-4674-b164-c141b817e4a0", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-175154447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36863339760945dc89398dad43684440", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap553edf3c-e4", "ovs_interfaceid": "553edf3c-e40b-4c94-b91c-f81a54df6643", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 936.817752] env[61857]: DEBUG oslo_concurrency.lockutils [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] Acquired lock "refresh_cache-825f5d87-76b1-4e4e-9b3d-4be419c7d323" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.818642] env[61857]: DEBUG nova.network.neutron [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Refreshing network info cache for port 553edf3c-e40b-4c94-b91c-f81a54df6643 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 936.823419] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:83:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '456bd8a2-0fb6-4b17-9d25-08e7995c5184', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '553edf3c-e40b-4c94-b91c-f81a54df6643', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.831801] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Creating folder: Project (36863339760945dc89398dad43684440). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 936.836016] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8183d314-2aba-4f0b-bbd7-b6e07bf5a809 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.850370] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Created folder: Project (36863339760945dc89398dad43684440) in parent group-v214027. [ 936.850370] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Creating folder: Instances. Parent ref: group-v214165. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 936.852844] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09192dee-4162-4e6f-b2fb-7d997f3b5668 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.864015] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Created folder: Instances in parent group-v214165. [ 936.864015] env[61857]: DEBUG oslo.service.loopingcall [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.864015] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 936.864015] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-163485c2-1a7b-4c46-9e05-c525bacfb389 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.892331] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951282, 'name': ReconfigVM_Task, 'duration_secs': 0.170263} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.894208] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 936.894697] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.894697] env[61857]: value = "task-951285" [ 936.894697] env[61857]: _type = "Task" [ 936.894697] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.895190] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b96ae889-5078-4581-9566-d518ed3d9a7d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.903389] env[61857]: DEBUG nova.compute.manager [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 936.910060] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 936.910060] env[61857]: value = "task-951286" [ 936.910060] env[61857]: _type = "Task" [ 936.910060] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.918848] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951285, 'name': CreateVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.924987] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951286, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.001134] env[61857]: INFO nova.compute.manager [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Rebuilding instance [ 937.006238] env[61857]: DEBUG nova.network.neutron [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updated VIF entry in instance network info cache for port 9d3974fc-a398-4220-9f8c-d78aa99796bc. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 937.006790] env[61857]: DEBUG nova.network.neutron [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.075110] env[61857]: DEBUG nova.compute.manager [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.076540] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae533b0-2ac2-43dc-8d2c-9919f9cb66f2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.109619] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39bdbce7-4c4b-41d3-819d-35319d8a11d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.117755] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf350be-a613-44a2-a640-e2e6aa53446f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.156102] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c48fa6e-dde5-48b9-9ec6-609f2d1fcea2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.165548] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b7056b-7fc5-4d2b-b0a0-1120d65a0b22 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.181525] env[61857]: DEBUG nova.compute.provider_tree [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.215602] env[61857]: DEBUG nova.network.neutron [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance_info_cache with network_info: [{"id": "e4c39a04-24d7-48d1-a089-2304778bd059", "address": "fa:16:3e:45:7a:cc", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4c39a04-24", "ovs_interfaceid": "e4c39a04-24d7-48d1-a089-2304778bd059", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.412804] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951285, 'name': CreateVM_Task, 'duration_secs': 0.496762} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.420186] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 937.421288] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.421536] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.421887] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 937.422728] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35253a08-d98f-43c8-be37-f5af6be95935 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.429212] env[61857]: DEBUG oslo_vmware.api [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951286, 'name': PowerOnVM_Task, 'duration_secs': 0.447435} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.430315] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 937.434510] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 937.434510] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52601ad9-225f-2a67-1945-db4ad83b8cdb" [ 937.434510] env[61857]: _type = "Task" [ 937.434510] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.436215] env[61857]: DEBUG nova.compute.manager [None req-f0b109d0-80f8-4f03-9363-3e4878183f1c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.437647] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6770cede-5915-4dd2-aca7-da2fb9542894 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.449464] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.464863] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52601ad9-225f-2a67-1945-db4ad83b8cdb, 'name': SearchDatastore_Task, 'duration_secs': 0.019048} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.465454] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.465694] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.466047] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.466252] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.466451] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.466727] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b302d15a-00be-4c4b-bb21-4e3e8a9fe326 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.476943] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.477178] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 937.480395] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3641160a-7d78-4912-aa12-88e196b330d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.486526] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 937.486526] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52370f0c-0cf9-04cb-e7c3-5448b815cd69" [ 937.486526] env[61857]: _type = "Task" [ 937.486526] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.496043] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52370f0c-0cf9-04cb-e7c3-5448b815cd69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.509416] env[61857]: DEBUG oslo_concurrency.lockutils [req-b3fbe447-25b8-45a7-a7ad-79298a4dcdb9 req-8c928663-cd0d-42c2-8fe4-f9b21e16f1cc service nova] Releasing lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.593140] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 937.593485] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d99b5f23-7411-4c70-abdb-d8e64a9701a4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.601066] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 937.601066] env[61857]: value = "task-951287" [ 937.601066] env[61857]: _type = "Task" [ 937.601066] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.607454] env[61857]: DEBUG nova.network.neutron [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Updated VIF entry in instance network info cache for port 553edf3c-e40b-4c94-b91c-f81a54df6643. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 937.607802] env[61857]: DEBUG nova.network.neutron [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Updating instance_info_cache with network_info: [{"id": "553edf3c-e40b-4c94-b91c-f81a54df6643", "address": "fa:16:3e:9b:83:91", "network": {"id": "60998eff-d613-4674-b164-c141b817e4a0", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-175154447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36863339760945dc89398dad43684440", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap553edf3c-e4", "ovs_interfaceid": "553edf3c-e40b-4c94-b91c-f81a54df6643", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.618218] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951287, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.683481] env[61857]: DEBUG nova.scheduler.client.report [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.718640] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.004316] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52370f0c-0cf9-04cb-e7c3-5448b815cd69, 'name': SearchDatastore_Task, 'duration_secs': 0.010614} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.004316] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9474a6d-1ae4-40e4-9673-fc095d898644 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.012461] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 938.012461] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e0b79c-96f7-33f1-ef12-fa714485efe5" [ 938.012461] env[61857]: _type = "Task" [ 938.012461] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.022951] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e0b79c-96f7-33f1-ef12-fa714485efe5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.110596] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951287, 'name': PowerOffVM_Task, 'duration_secs': 0.338524} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.111337] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 938.111337] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 938.112042] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712b928a-4f30-4b4f-9cd3-8674f5d66864 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.116932] env[61857]: DEBUG oslo_concurrency.lockutils [req-3c92e897-bc6c-4b4f-821c-859db5923e0b req-649b8a8e-8ef9-4a20-941a-12dea5c5fe71 service nova] Releasing lock "refresh_cache-825f5d87-76b1-4e4e-9b3d-4be419c7d323" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.122104] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 938.122104] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-acdbfdb3-31e5-47bf-b5f4-58353f2f2a9e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.190878] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.190878] env[61857]: DEBUG nova.compute.manager [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 938.192612] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.018s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.194430] env[61857]: INFO nova.compute.claims [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 938.199956] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 938.200495] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 938.200855] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleting the datastore file [datastore2] d85f8a2d-6053-45ec-81bb-de9fe02af3fe {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.201454] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14ee9af4-b724-4822-b791-d6b983992d63 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.209629] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 938.209629] env[61857]: value = "task-951289" [ 938.209629] env[61857]: _type = "Task" [ 938.209629] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.232742] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951289, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.524271] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e0b79c-96f7-33f1-ef12-fa714485efe5, 'name': SearchDatastore_Task, 'duration_secs': 0.010304} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.524573] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.524839] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 825f5d87-76b1-4e4e-9b3d-4be419c7d323/825f5d87-76b1-4e4e-9b3d-4be419c7d323.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 938.525424] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5b03246-aa82-47c3-81b5-4ec5fa7f6c8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.531644] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 938.531644] env[61857]: value = "task-951290" [ 938.531644] env[61857]: _type = "Task" [ 938.531644] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.541445] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951290, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.701180] env[61857]: DEBUG nova.compute.utils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 938.703264] env[61857]: DEBUG nova.compute.manager [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 938.703572] env[61857]: DEBUG nova.network.neutron [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 938.724644] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951289, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.378334} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.725066] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.725384] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 938.725701] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 938.820235] env[61857]: DEBUG nova.policy [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1496582f8474b14baf63b215a80e64e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3688a3ca85044527a84ad1a7b201fbdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 938.917992] env[61857]: INFO nova.compute.manager [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Unrescuing [ 938.918435] env[61857]: DEBUG oslo_concurrency.lockutils [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.918632] env[61857]: DEBUG oslo_concurrency.lockutils [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquired lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.918899] env[61857]: DEBUG nova.network.neutron [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 939.042801] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951290, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.210959] env[61857]: DEBUG nova.compute.manager [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 939.241031] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bb779a-1276-42f9-87b0-735ec2c6e634 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.276734] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance '18a632be-322f-48a9-becf-51ff3b735ad4' progress to 0 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 939.378511] env[61857]: DEBUG nova.network.neutron [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Successfully created port: d64de57d-2c7b-4898-8fc2-0ebf0bc06faf {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.516525] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98563713-2373-4409-824b-bc3e35dc144a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.524425] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6507e397-8bc1-42b5-8582-6d6e212c28f0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.564616] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b7303c-69ca-4ed1-8704-fbb81da92d5e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.573184] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951290, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.658953} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.575323] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 825f5d87-76b1-4e4e-9b3d-4be419c7d323/825f5d87-76b1-4e4e-9b3d-4be419c7d323.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 939.575566] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 939.576555] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f903ced8-e17b-43db-a1c0-1b7184db433a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.579198] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b38ebf5e-4e2c-4f9b-951a-d556ade6cf4f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.593883] env[61857]: DEBUG nova.compute.provider_tree [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.597162] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 939.597162] env[61857]: value = "task-951291" [ 939.597162] env[61857]: _type = "Task" [ 939.597162] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.605949] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951291, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.770020] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.770498] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.770977] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.771104] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.771282] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.771469] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.772424] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.772801] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.772952] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.773152] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.773436] env[61857]: DEBUG nova.virt.hardware [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.774553] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659bf2e7-8f9a-48a6-b022-f7ea9154d485 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.782907] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 939.783595] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac3527fc-85bc-48cc-8388-7925b6f03c1e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.786506] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9f3ea5-e4fa-4e13-af97-a2285a5c0b7a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.806715] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:6f:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b47d366-f4ff-4148-a2b4-be60b9228074', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 939.815027] env[61857]: DEBUG oslo.service.loopingcall [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.817241] env[61857]: DEBUG nova.network.neutron [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updating instance_info_cache with network_info: [{"id": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "address": "fa:16:3e:94:19:6b", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9f80f94-9b", "ovs_interfaceid": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.818512] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 939.818877] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 939.818877] env[61857]: value = "task-951292" [ 939.818877] env[61857]: _type = "Task" [ 939.818877] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.819375] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a8bf28fa-6574-4c52-95e6-2bd9fe679fb4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.849366] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951292, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.851794] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 939.851794] env[61857]: value = "task-951293" [ 939.851794] env[61857]: _type = "Task" [ 939.851794] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.860209] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951293, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.097150] env[61857]: DEBUG nova.scheduler.client.report [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.110899] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951291, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071787} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.111251] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 940.112247] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f62cbf-a19a-4688-8780-b9a867374317 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.134982] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 825f5d87-76b1-4e4e-9b3d-4be419c7d323/825f5d87-76b1-4e4e-9b3d-4be419c7d323.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.135858] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3e78fe6-5380-4ac0-a63f-baf7a94909db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.158404] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 940.158404] env[61857]: value = "task-951294" [ 940.158404] env[61857]: _type = "Task" [ 940.158404] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.167136] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951294, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.224179] env[61857]: DEBUG nova.compute.manager [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.271066] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.271360] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.271661] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.271989] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.272272] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.272589] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.273203] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.273297] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.273774] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.273860] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.274643] env[61857]: DEBUG nova.virt.hardware [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.276937] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a5e0c6-b88a-45ee-8a8b-9eb5a799a6a3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.294203] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d6f86e-0320-42fa-b50a-66e584f764a5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.320365] env[61857]: DEBUG oslo_concurrency.lockutils [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Releasing lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.321243] env[61857]: DEBUG nova.objects.instance [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lazy-loading 'flavor' on Instance uuid d9d05629-79ef-45c6-ac54-b3bc18d306f4 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.350339] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951292, 'name': PowerOffVM_Task, 'duration_secs': 0.429382} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.350728] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 940.351017] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance '18a632be-322f-48a9-becf-51ff3b735ad4' progress to 17 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 940.370848] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951293, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.608516] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.609116] env[61857]: DEBUG nova.compute.manager [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 940.612077] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.079s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.612195] env[61857]: DEBUG nova.objects.instance [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lazy-loading 'resources' on Instance uuid e5e64477-a1aa-4a4d-91a2-b17d912e09c3 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.668069] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951294, 'name': ReconfigVM_Task, 'duration_secs': 0.428022} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.668406] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 825f5d87-76b1-4e4e-9b3d-4be419c7d323/825f5d87-76b1-4e4e-9b3d-4be419c7d323.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.669190] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9221b467-ec51-4c59-827b-136276ef56f1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.675235] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 940.675235] env[61857]: value = "task-951295" [ 940.675235] env[61857]: _type = "Task" [ 940.675235] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.684592] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951295, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.834098] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94ff126-21b8-4006-911b-2a37151fec30 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.864888] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.865177] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.865388] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.865660] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.866312] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.866312] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.866312] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.867656] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.867656] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.867853] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.867924] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.873420] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 940.879231] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d31c0795-f2ed-4920-bda1-bd9c80a43ce5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.889380] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84ccadae-b3e9-47e8-b92d-046de4fdf2db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.897403] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951293, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.899837] env[61857]: DEBUG oslo_vmware.api [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 940.899837] env[61857]: value = "task-951296" [ 940.899837] env[61857]: _type = "Task" [ 940.899837] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.900386] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 940.900386] env[61857]: value = "task-951297" [ 940.900386] env[61857]: _type = "Task" [ 940.900386] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.917980] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951297, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.921655] env[61857]: DEBUG oslo_vmware.api [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.967089] env[61857]: DEBUG nova.compute.manager [req-08584f84-a045-4823-9a6c-27dd155b9705 req-218dee4c-a50d-44c8-bde4-5cb0e44844d5 service nova] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Received event network-vif-plugged-d64de57d-2c7b-4898-8fc2-0ebf0bc06faf {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 940.967494] env[61857]: DEBUG oslo_concurrency.lockutils [req-08584f84-a045-4823-9a6c-27dd155b9705 req-218dee4c-a50d-44c8-bde4-5cb0e44844d5 service nova] Acquiring lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.967859] env[61857]: DEBUG oslo_concurrency.lockutils [req-08584f84-a045-4823-9a6c-27dd155b9705 req-218dee4c-a50d-44c8-bde4-5cb0e44844d5 service nova] Lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.968406] env[61857]: DEBUG oslo_concurrency.lockutils [req-08584f84-a045-4823-9a6c-27dd155b9705 req-218dee4c-a50d-44c8-bde4-5cb0e44844d5 service nova] Lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.968501] env[61857]: DEBUG nova.compute.manager [req-08584f84-a045-4823-9a6c-27dd155b9705 req-218dee4c-a50d-44c8-bde4-5cb0e44844d5 service nova] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] No waiting events found dispatching network-vif-plugged-d64de57d-2c7b-4898-8fc2-0ebf0bc06faf {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 940.968992] env[61857]: WARNING nova.compute.manager [req-08584f84-a045-4823-9a6c-27dd155b9705 req-218dee4c-a50d-44c8-bde4-5cb0e44844d5 service nova] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Received unexpected event network-vif-plugged-d64de57d-2c7b-4898-8fc2-0ebf0bc06faf for instance with vm_state building and task_state spawning. [ 941.117172] env[61857]: DEBUG nova.compute.utils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 941.117840] env[61857]: DEBUG nova.compute.manager [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 941.118176] env[61857]: DEBUG nova.network.neutron [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 941.157064] env[61857]: DEBUG nova.network.neutron [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Successfully updated port: d64de57d-2c7b-4898-8fc2-0ebf0bc06faf {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.180803] env[61857]: DEBUG nova.policy [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '968509a1e5d04164aa81abed41aa845f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8dc7382605d14c748e938080baf494ca', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 941.193056] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951295, 'name': Rename_Task, 'duration_secs': 0.14684} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.193458] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 941.193736] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99618e3d-fb6a-446e-940e-c0c029e9fd8e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.207021] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 941.207021] env[61857]: value = "task-951298" [ 941.207021] env[61857]: _type = "Task" [ 941.207021] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.212312] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951298, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.372814] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951293, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.419258] env[61857]: DEBUG oslo_vmware.api [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951296, 'name': PowerOffVM_Task, 'duration_secs': 0.300672} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.427637] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 941.438309] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Reconfiguring VM instance instance-0000004d to detach disk 2002 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 941.438309] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951297, 'name': ReconfigVM_Task, 'duration_secs': 0.208364} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.438309] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-501cb99a-274c-44fc-817a-32918720d7e3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.451513] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance '18a632be-322f-48a9-becf-51ff3b735ad4' progress to 33 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 941.458146] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d73920-7fb6-4755-b29e-9c6dc94387fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.464522] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f55650e-a8c3-426f-8ec5-eae3b5dc08d5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.468177] env[61857]: DEBUG oslo_vmware.api [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 941.468177] env[61857]: value = "task-951299" [ 941.468177] env[61857]: _type = "Task" [ 941.468177] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.503355] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee11d6fe-c77e-4c9d-8a89-19cebef532ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.510324] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquiring lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.510596] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.511855] env[61857]: DEBUG oslo_vmware.api [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951299, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.517140] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8a9e20-47f4-4c34-89a3-afb25e300905 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.531658] env[61857]: DEBUG nova.compute.provider_tree [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.599712] env[61857]: DEBUG nova.network.neutron [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Successfully created port: a5b81295-b253-46a2-b1a7-78d459f101de {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.624277] env[61857]: DEBUG nova.compute.manager [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 941.660969] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "refresh_cache-ce5f713d-8b63-4b95-a5ad-feddb95f921d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.660969] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "refresh_cache-ce5f713d-8b63-4b95-a5ad-feddb95f921d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.660969] env[61857]: DEBUG nova.network.neutron [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 941.715958] env[61857]: DEBUG oslo_vmware.api [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951298, 'name': PowerOnVM_Task, 'duration_secs': 0.454737} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.718201] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 941.718201] env[61857]: INFO nova.compute.manager [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Took 8.30 seconds to spawn the instance on the hypervisor. [ 941.718201] env[61857]: DEBUG nova.compute.manager [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.718201] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7370ee7e-2160-4e74-a4b7-eda724992664 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.775024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.775024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.874737] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951293, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.961312] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.961589] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.961739] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.962808] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.962808] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.962808] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.962808] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.963044] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.963044] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.963260] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.964050] env[61857]: DEBUG nova.virt.hardware [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.968577] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Reconfiguring VM instance instance-00000052 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 941.968928] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c40adedf-c0bd-4e74-b6cc-9ba289b9e2bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.991502] env[61857]: DEBUG oslo_vmware.api [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951299, 'name': ReconfigVM_Task, 'duration_secs': 0.279316} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.993268] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Reconfigured VM instance instance-0000004d to detach disk 2002 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 941.993465] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 941.993810] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 941.993810] env[61857]: value = "task-951300" [ 941.993810] env[61857]: _type = "Task" [ 941.993810] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.994087] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f3934ef-bfc2-4235-88fa-747c71d52d7e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.006144] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951300, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.007645] env[61857]: DEBUG oslo_vmware.api [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 942.007645] env[61857]: value = "task-951301" [ 942.007645] env[61857]: _type = "Task" [ 942.007645] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.015333] env[61857]: DEBUG nova.compute.manager [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 942.018359] env[61857]: DEBUG oslo_vmware.api [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951301, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.035779] env[61857]: DEBUG nova.scheduler.client.report [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.193293] env[61857]: DEBUG nova.network.neutron [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 942.237213] env[61857]: INFO nova.compute.manager [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Took 28.76 seconds to build instance. [ 942.277625] env[61857]: DEBUG nova.compute.utils [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.342086] env[61857]: DEBUG nova.network.neutron [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Updating instance_info_cache with network_info: [{"id": "d64de57d-2c7b-4898-8fc2-0ebf0bc06faf", "address": "fa:16:3e:6a:29:63", "network": {"id": "e7b01bbf-aac5-42cb-b166-da2a3a70a175", "bridge": "br-int", "label": "tempest-ImagesTestJSON-229290731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3688a3ca85044527a84ad1a7b201fbdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd64de57d-2c", "ovs_interfaceid": "d64de57d-2c7b-4898-8fc2-0ebf0bc06faf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.373559] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951293, 'name': CreateVM_Task, 'duration_secs': 2.23287} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.373749] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 942.374464] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.374639] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.374969] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.375244] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48bf7fed-8f93-4c56-93b2-d6c17076eb4c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.380067] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 942.380067] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52594d11-447f-7201-ac33-57685fc99bca" [ 942.380067] env[61857]: _type = "Task" [ 942.380067] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.388494] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52594d11-447f-7201-ac33-57685fc99bca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.505165] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951300, 'name': ReconfigVM_Task, 'duration_secs': 0.196576} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.505448] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Reconfigured VM instance instance-00000052 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 942.506208] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8ce78a-6c32-4a1e-85d5-40a57552d529 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.516201] env[61857]: DEBUG oslo_vmware.api [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951301, 'name': PowerOnVM_Task, 'duration_secs': 0.419944} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.526980] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 942.526980] env[61857]: DEBUG nova.compute.manager [None req-db6299f7-478a-4b8c-8bb8-583c635fed50 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.539691] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 18a632be-322f-48a9-becf-51ff3b735ad4/18a632be-322f-48a9-becf-51ff3b735ad4.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.540760] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6cf27b-12f4-49ca-85ff-2420b9b071c9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.543344] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff3f04f1-e58f-4840-9bab-380787d42682 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.558535] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.559276] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.947s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.561412] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.811s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.562883] env[61857]: INFO nova.compute.claims [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.574082] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 942.574082] env[61857]: value = "task-951302" [ 942.574082] env[61857]: _type = "Task" [ 942.574082] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.579265] env[61857]: INFO nova.scheduler.client.report [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleted allocations for instance e5e64477-a1aa-4a4d-91a2-b17d912e09c3 [ 942.584113] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951302, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.634840] env[61857]: DEBUG nova.compute.manager [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 942.664134] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.664223] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.664403] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.664688] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.664828] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.665087] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.665295] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.665504] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.665713] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.665929] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.666204] env[61857]: DEBUG nova.virt.hardware [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.667462] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71fd3479-37d8-4342-8fd9-57ff242aac10 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.676649] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2390d2d9-08b1-462b-943c-0a5da6a92f05 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.742851] env[61857]: DEBUG oslo_concurrency.lockutils [None req-aa409114-8aec-4751-b61b-ff99d666ce1b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.275s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.780289] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.845565] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "refresh_cache-ce5f713d-8b63-4b95-a5ad-feddb95f921d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.845565] env[61857]: DEBUG nova.compute.manager [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Instance network_info: |[{"id": "d64de57d-2c7b-4898-8fc2-0ebf0bc06faf", "address": "fa:16:3e:6a:29:63", "network": {"id": "e7b01bbf-aac5-42cb-b166-da2a3a70a175", "bridge": "br-int", "label": "tempest-ImagesTestJSON-229290731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3688a3ca85044527a84ad1a7b201fbdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd64de57d-2c", "ovs_interfaceid": "d64de57d-2c7b-4898-8fc2-0ebf0bc06faf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.845981] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:29:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4576b9d4-535c-40aa-b078-246f671f216e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd64de57d-2c7b-4898-8fc2-0ebf0bc06faf', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.853751] env[61857]: DEBUG oslo.service.loopingcall [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.853994] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 942.854213] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa4a318a-a613-4b83-9a04-4708f6bc4bba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.874474] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.874474] env[61857]: value = "task-951303" [ 942.874474] env[61857]: _type = "Task" [ 942.874474] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.882317] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951303, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.890693] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52594d11-447f-7201-ac33-57685fc99bca, 'name': SearchDatastore_Task, 'duration_secs': 0.014716} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.891011] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.891292] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.891559] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.891753] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.891965] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.892264] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f116c4ef-4213-40f9-88de-59faec1d6acf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.900887] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.901077] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 942.902150] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f07be66-3bae-474c-8d33-bf69c9891d80 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.908570] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 942.908570] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5263f4cd-88b1-b0c6-0ee1-16583205a4c3" [ 942.908570] env[61857]: _type = "Task" [ 942.908570] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.917476] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5263f4cd-88b1-b0c6-0ee1-16583205a4c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.098375] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951302, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.100273] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a41655ea-40ee-497d-824e-35d88200172c tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "e5e64477-a1aa-4a4d-91a2-b17d912e09c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.679s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.165264] env[61857]: DEBUG nova.network.neutron [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Successfully updated port: a5b81295-b253-46a2-b1a7-78d459f101de {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.362987] env[61857]: DEBUG nova.compute.manager [req-79350340-bff0-450a-92a2-2c8968dbefe5 req-3eef27a2-6829-4893-b49d-dfca7c3ee698 service nova] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Received event network-changed-d64de57d-2c7b-4898-8fc2-0ebf0bc06faf {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 943.363221] env[61857]: DEBUG nova.compute.manager [req-79350340-bff0-450a-92a2-2c8968dbefe5 req-3eef27a2-6829-4893-b49d-dfca7c3ee698 service nova] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Refreshing instance network info cache due to event network-changed-d64de57d-2c7b-4898-8fc2-0ebf0bc06faf. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 943.363446] env[61857]: DEBUG oslo_concurrency.lockutils [req-79350340-bff0-450a-92a2-2c8968dbefe5 req-3eef27a2-6829-4893-b49d-dfca7c3ee698 service nova] Acquiring lock "refresh_cache-ce5f713d-8b63-4b95-a5ad-feddb95f921d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.363615] env[61857]: DEBUG oslo_concurrency.lockutils [req-79350340-bff0-450a-92a2-2c8968dbefe5 req-3eef27a2-6829-4893-b49d-dfca7c3ee698 service nova] Acquired lock "refresh_cache-ce5f713d-8b63-4b95-a5ad-feddb95f921d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.363775] env[61857]: DEBUG nova.network.neutron [req-79350340-bff0-450a-92a2-2c8968dbefe5 req-3eef27a2-6829-4893-b49d-dfca7c3ee698 service nova] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Refreshing network info cache for port d64de57d-2c7b-4898-8fc2-0ebf0bc06faf {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 943.384487] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951303, 'name': CreateVM_Task, 'duration_secs': 0.327638} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.384686] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 943.385365] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.385538] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.385849] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.386117] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-760cf209-8749-402e-a968-0f30ba841982 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.390556] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 943.390556] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525f6e2d-a920-0ced-0342-c2f8cb8859e9" [ 943.390556] env[61857]: _type = "Task" [ 943.390556] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.401151] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525f6e2d-a920-0ced-0342-c2f8cb8859e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.418142] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5263f4cd-88b1-b0c6-0ee1-16583205a4c3, 'name': SearchDatastore_Task, 'duration_secs': 0.013254} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.418853] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1104577-5d22-4e6e-a0c3-8d28c74fa567 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.423599] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 943.423599] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ff3c47-2c72-faf2-2380-eec7d7e99a64" [ 943.423599] env[61857]: _type = "Task" [ 943.423599] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.431041] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ff3c47-2c72-faf2-2380-eec7d7e99a64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.587420] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951302, 'name': ReconfigVM_Task, 'duration_secs': 0.567868} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.588086] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 18a632be-322f-48a9-becf-51ff3b735ad4/18a632be-322f-48a9-becf-51ff3b735ad4.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.588357] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance '18a632be-322f-48a9-becf-51ff3b735ad4' progress to 50 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 943.667415] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquiring lock "refresh_cache-c821ec43-9462-4276-bdc9-d54e64c0a3cf" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.667728] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquired lock "refresh_cache-c821ec43-9462-4276-bdc9-d54e64c0a3cf" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.667728] env[61857]: DEBUG nova.network.neutron [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 943.815213] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b750ee-436c-468b-a72c-4313d40eae0e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.823329] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407e66e4-36e4-400f-a4ec-5c35e8f2bea3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.853115] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.853372] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.853603] env[61857]: INFO nova.compute.manager [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Attaching volume e2d3aded-f562-44fa-b5fc-9f83f20bc32d to /dev/sdb [ 943.855948] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4a214e-ac71-472b-9af4-abadaf6be898 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.870259] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6cf2df-c67e-42f5-9e08-aa8335f56cc5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.885236] env[61857]: DEBUG nova.compute.provider_tree [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.892345] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a02b43-a824-4759-9774-dfa60f052a01 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.907317] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4539c84d-077f-40aa-b09c-63ca52752ed6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.909967] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525f6e2d-a920-0ced-0342-c2f8cb8859e9, 'name': SearchDatastore_Task, 'duration_secs': 0.012707} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.910596] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.910884] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.911146] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.922334] env[61857]: DEBUG nova.virt.block_device [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Updating existing volume attachment record: 22898bd8-660f-4f37-a2c9-ba57ed52a2c5 {{(pid=61857) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 943.932068] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ff3c47-2c72-faf2-2380-eec7d7e99a64, 'name': SearchDatastore_Task, 'duration_secs': 0.045861} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.932326] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.932577] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] d85f8a2d-6053-45ec-81bb-de9fe02af3fe/d85f8a2d-6053-45ec-81bb-de9fe02af3fe.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 943.932850] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.933164] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.933261] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d90d873e-9481-4e0d-98c4-6beb84f91d49 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.934988] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9594da34-12e4-49b9-8480-4c99c0b2948e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.941557] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 943.941557] env[61857]: value = "task-951304" [ 943.941557] env[61857]: _type = "Task" [ 943.941557] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.947339] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.947520] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 943.948536] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18d52e6f-cb73-4906-914c-ce1b338b68d1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.953953] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951304, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.956814] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 943.956814] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad8bb7-3d5d-69d0-1423-973e291b6769" [ 943.956814] env[61857]: _type = "Task" [ 943.956814] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.964213] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad8bb7-3d5d-69d0-1423-973e291b6769, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.024402] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquiring lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.024640] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.024877] env[61857]: INFO nova.compute.manager [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Rebooting instance [ 944.099109] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-500aa513-15d7-4235-87b4-8a6ceb07acb1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.103101] env[61857]: DEBUG nova.compute.manager [req-0ef5f22f-6247-450f-ac2e-592c4ef00ac9 req-ebedd2cb-9a08-4fd1-8fb1-4da8d788b114 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Received event network-changed-d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 944.103312] env[61857]: DEBUG nova.compute.manager [req-0ef5f22f-6247-450f-ac2e-592c4ef00ac9 req-ebedd2cb-9a08-4fd1-8fb1-4da8d788b114 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Refreshing instance network info cache due to event network-changed-d9f80f94-9be0-4d62-863f-7562c6c9e63e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 944.103536] env[61857]: DEBUG oslo_concurrency.lockutils [req-0ef5f22f-6247-450f-ac2e-592c4ef00ac9 req-ebedd2cb-9a08-4fd1-8fb1-4da8d788b114 service nova] Acquiring lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.103687] env[61857]: DEBUG oslo_concurrency.lockutils [req-0ef5f22f-6247-450f-ac2e-592c4ef00ac9 req-ebedd2cb-9a08-4fd1-8fb1-4da8d788b114 service nova] Acquired lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.103851] env[61857]: DEBUG nova.network.neutron [req-0ef5f22f-6247-450f-ac2e-592c4ef00ac9 req-ebedd2cb-9a08-4fd1-8fb1-4da8d788b114 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Refreshing network info cache for port d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 944.122300] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd01c28-4ffd-4b60-b1f8-031060ce6eb5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.125660] env[61857]: DEBUG nova.network.neutron [req-79350340-bff0-450a-92a2-2c8968dbefe5 req-3eef27a2-6829-4893-b49d-dfca7c3ee698 service nova] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Updated VIF entry in instance network info cache for port d64de57d-2c7b-4898-8fc2-0ebf0bc06faf. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 944.126087] env[61857]: DEBUG nova.network.neutron [req-79350340-bff0-450a-92a2-2c8968dbefe5 req-3eef27a2-6829-4893-b49d-dfca7c3ee698 service nova] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Updating instance_info_cache with network_info: [{"id": "d64de57d-2c7b-4898-8fc2-0ebf0bc06faf", "address": "fa:16:3e:6a:29:63", "network": {"id": "e7b01bbf-aac5-42cb-b166-da2a3a70a175", "bridge": "br-int", "label": "tempest-ImagesTestJSON-229290731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3688a3ca85044527a84ad1a7b201fbdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4576b9d4-535c-40aa-b078-246f671f216e", "external-id": "nsx-vlan-transportzone-27", "segmentation_id": 27, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd64de57d-2c", "ovs_interfaceid": "d64de57d-2c7b-4898-8fc2-0ebf0bc06faf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.150710] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance '18a632be-322f-48a9-becf-51ff3b735ad4' progress to 67 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 944.239269] env[61857]: DEBUG nova.network.neutron [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 944.389262] env[61857]: DEBUG nova.scheduler.client.report [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.452535] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951304, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.463182] env[61857]: DEBUG nova.network.neutron [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Updating instance_info_cache with network_info: [{"id": "a5b81295-b253-46a2-b1a7-78d459f101de", "address": "fa:16:3e:59:26:5d", "network": {"id": "c3318ddc-c0fb-491b-badf-007720f0c9e1", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1428298316-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8dc7382605d14c748e938080baf494ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5b81295-b2", "ovs_interfaceid": "a5b81295-b253-46a2-b1a7-78d459f101de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.470663] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad8bb7-3d5d-69d0-1423-973e291b6769, 'name': SearchDatastore_Task, 'duration_secs': 0.012423} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.471581] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad4bbb2c-fd33-4d90-b65a-88d9729cfa77 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.477223] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 944.477223] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5229db9d-aec4-acd4-0b1d-ae6ec0310e7e" [ 944.477223] env[61857]: _type = "Task" [ 944.477223] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.485541] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5229db9d-aec4-acd4-0b1d-ae6ec0310e7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.549118] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquiring lock "refresh_cache-825f5d87-76b1-4e4e-9b3d-4be419c7d323" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.549448] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquired lock "refresh_cache-825f5d87-76b1-4e4e-9b3d-4be419c7d323" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.550325] env[61857]: DEBUG nova.network.neutron [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 944.629603] env[61857]: DEBUG oslo_concurrency.lockutils [req-79350340-bff0-450a-92a2-2c8968dbefe5 req-3eef27a2-6829-4893-b49d-dfca7c3ee698 service nova] Releasing lock "refresh_cache-ce5f713d-8b63-4b95-a5ad-feddb95f921d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.754971] env[61857]: DEBUG nova.network.neutron [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Port e4c39a04-24d7-48d1-a089-2304778bd059 binding to destination host cpu-1 is already ACTIVE {{(pid=61857) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 944.896970] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.897478] env[61857]: DEBUG nova.compute.manager [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 944.907918] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.756s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.957680] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951304, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.529843} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.957976] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] d85f8a2d-6053-45ec-81bb-de9fe02af3fe/d85f8a2d-6053-45ec-81bb-de9fe02af3fe.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 944.958275] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.958543] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-17f6b4a3-988d-47cc-8994-bb587d48b22e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.965567] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Releasing lock "refresh_cache-c821ec43-9462-4276-bdc9-d54e64c0a3cf" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.965852] env[61857]: DEBUG nova.compute.manager [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Instance network_info: |[{"id": "a5b81295-b253-46a2-b1a7-78d459f101de", "address": "fa:16:3e:59:26:5d", "network": {"id": "c3318ddc-c0fb-491b-badf-007720f0c9e1", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1428298316-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8dc7382605d14c748e938080baf494ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5b81295-b2", "ovs_interfaceid": "a5b81295-b253-46a2-b1a7-78d459f101de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 944.966194] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 944.966194] env[61857]: value = "task-951306" [ 944.966194] env[61857]: _type = "Task" [ 944.966194] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.966611] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:26:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32faf59b-014c-4f1f-8331-40df95bf741f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5b81295-b253-46a2-b1a7-78d459f101de', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 944.974695] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Creating folder: Project (8dc7382605d14c748e938080baf494ca). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 944.975855] env[61857]: DEBUG nova.network.neutron [req-0ef5f22f-6247-450f-ac2e-592c4ef00ac9 req-ebedd2cb-9a08-4fd1-8fb1-4da8d788b114 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updated VIF entry in instance network info cache for port d9f80f94-9be0-4d62-863f-7562c6c9e63e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 944.976202] env[61857]: DEBUG nova.network.neutron [req-0ef5f22f-6247-450f-ac2e-592c4ef00ac9 req-ebedd2cb-9a08-4fd1-8fb1-4da8d788b114 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updating instance_info_cache with network_info: [{"id": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "address": "fa:16:3e:94:19:6b", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9f80f94-9b", "ovs_interfaceid": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.978650] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-515991ce-dcb4-45a8-aee7-3eb2d8b7c103 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.991790] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951306, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.995397] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5229db9d-aec4-acd4-0b1d-ae6ec0310e7e, 'name': SearchDatastore_Task, 'duration_secs': 0.008756} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.996546] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.996813] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] ce5f713d-8b63-4b95-a5ad-feddb95f921d/ce5f713d-8b63-4b95-a5ad-feddb95f921d.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 944.997093] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Created folder: Project (8dc7382605d14c748e938080baf494ca) in parent group-v214027. [ 944.997269] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Creating folder: Instances. Parent ref: group-v214171. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 944.997491] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e690b6f-ba04-4f57-90e1-753cdb00b728 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.999302] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ac76135-0daa-4005-bd8a-bd1eca65a602 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.004570] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 945.004570] env[61857]: value = "task-951308" [ 945.004570] env[61857]: _type = "Task" [ 945.004570] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.009961] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Created folder: Instances in parent group-v214171. [ 945.010255] env[61857]: DEBUG oslo.service.loopingcall [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.013577] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 945.013577] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.013725] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1fcf725-b356-4944-92d9-f78c6841be61 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.034330] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.034330] env[61857]: value = "task-951310" [ 945.034330] env[61857]: _type = "Task" [ 945.034330] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.042193] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951310, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.265467] env[61857]: DEBUG nova.network.neutron [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Updating instance_info_cache with network_info: [{"id": "553edf3c-e40b-4c94-b91c-f81a54df6643", "address": "fa:16:3e:9b:83:91", "network": {"id": "60998eff-d613-4674-b164-c141b817e4a0", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-175154447-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36863339760945dc89398dad43684440", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap553edf3c-e4", "ovs_interfaceid": "553edf3c-e40b-4c94-b91c-f81a54df6643", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.389619] env[61857]: DEBUG nova.compute.manager [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Received event network-vif-plugged-a5b81295-b253-46a2-b1a7-78d459f101de {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.389814] env[61857]: DEBUG oslo_concurrency.lockutils [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] Acquiring lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.390143] env[61857]: DEBUG oslo_concurrency.lockutils [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] Lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.390427] env[61857]: DEBUG oslo_concurrency.lockutils [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] Lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.390923] env[61857]: DEBUG nova.compute.manager [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] No waiting events found dispatching network-vif-plugged-a5b81295-b253-46a2-b1a7-78d459f101de {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 945.391067] env[61857]: WARNING nova.compute.manager [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Received unexpected event network-vif-plugged-a5b81295-b253-46a2-b1a7-78d459f101de for instance with vm_state building and task_state spawning. [ 945.391277] env[61857]: DEBUG nova.compute.manager [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Received event network-changed-a5b81295-b253-46a2-b1a7-78d459f101de {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.391537] env[61857]: DEBUG nova.compute.manager [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Refreshing instance network info cache due to event network-changed-a5b81295-b253-46a2-b1a7-78d459f101de. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 945.391807] env[61857]: DEBUG oslo_concurrency.lockutils [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] Acquiring lock "refresh_cache-c821ec43-9462-4276-bdc9-d54e64c0a3cf" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.392053] env[61857]: DEBUG oslo_concurrency.lockutils [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] Acquired lock "refresh_cache-c821ec43-9462-4276-bdc9-d54e64c0a3cf" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.392636] env[61857]: DEBUG nova.network.neutron [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Refreshing network info cache for port a5b81295-b253-46a2-b1a7-78d459f101de {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 945.415497] env[61857]: DEBUG nova.compute.utils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 945.417568] env[61857]: DEBUG nova.compute.manager [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 945.417726] env[61857]: DEBUG nova.network.neutron [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 945.469464] env[61857]: DEBUG nova.policy [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98001d3024e44cb2a9cee667dac33574', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35ae3038a282412e963845e221e87f67', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 945.485768] env[61857]: DEBUG oslo_concurrency.lockutils [req-0ef5f22f-6247-450f-ac2e-592c4ef00ac9 req-ebedd2cb-9a08-4fd1-8fb1-4da8d788b114 service nova] Releasing lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.486238] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951306, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.282996} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.486494] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.487367] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c69f67-56d8-4146-9fc2-66bb50afbf63 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.510979] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] d85f8a2d-6053-45ec-81bb-de9fe02af3fe/d85f8a2d-6053-45ec-81bb-de9fe02af3fe.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.511350] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-191db623-554f-46f3-a165-04b9e498bf8d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.535157] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.538813] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 945.538813] env[61857]: value = "task-951311" [ 945.538813] env[61857]: _type = "Task" [ 945.538813] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.544769] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951310, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.549630] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951311, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.780859] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Releasing lock "refresh_cache-825f5d87-76b1-4e4e-9b3d-4be419c7d323" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.797673] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "18a632be-322f-48a9-becf-51ff3b735ad4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.800027] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "18a632be-322f-48a9-becf-51ff3b735ad4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.800027] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "18a632be-322f-48a9-becf-51ff3b735ad4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.801867] env[61857]: DEBUG nova.compute.manager [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.804695] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f2856c-3784-4e4b-9007-b5aca0631358 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.860605] env[61857]: DEBUG nova.network.neutron [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Successfully created port: 507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 945.918270] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Applying migration context for instance 18a632be-322f-48a9-becf-51ff3b735ad4 as it has an incoming, in-progress migration c72d05c4-0844-486a-bf03-11222909c20c. Migration status is post-migrating {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 945.919993] env[61857]: INFO nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating resource usage from migration c72d05c4-0844-486a-bf03-11222909c20c [ 945.922775] env[61857]: DEBUG nova.compute.manager [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 945.952027] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 18d9307d-6a20-47b4-b4e9-176f9a8c33cc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.952249] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 71920a62-090d-4df1-937e-87df3b043e28 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.952388] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 28bf59ca-4ffe-4005-9a88-da0660ebb48a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.952511] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e91e99b9-1cd3-4345-af09-f14af4df1214 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.952626] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance d9d05629-79ef-45c6-ac54-b3bc18d306f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.952782] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 0de8e67e-24b9-48d0-ac90-b5ec5b93e243 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.952942] env[61857]: WARNING nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 992bcab3-7396-495f-96a1-e99ecd13c961 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 945.953080] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance f982f0c5-c701-4e5b-b96d-c21a3c0da24f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.953199] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance d85f8a2d-6053-45ec-81bb-de9fe02af3fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.953349] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 825f5d87-76b1-4e4e-9b3d-4be419c7d323 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.953484] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Migration c72d05c4-0844-486a-bf03-11222909c20c is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 945.953600] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 18a632be-322f-48a9-becf-51ff3b735ad4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.954071] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance ce5f713d-8b63-4b95-a5ad-feddb95f921d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.954071] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c821ec43-9462-4276-bdc9-d54e64c0a3cf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.954071] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance a42e8c77-3cc3-4323-b1ee-2a95192910d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 945.954071] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (2d5860d0-8574-4e55-9ced-20e66f0314c2): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 946.023877] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951308, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.050534] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951310, 'name': CreateVM_Task, 'duration_secs': 0.785245} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.054632] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 946.055043] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.055762] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.055923] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.056273] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 946.056575] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e54601d2-48ec-4fd4-930e-06e8b57621ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.062090] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for the task: (returnval){ [ 946.062090] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52abe30b-86c4-0a36-60fd-47823fcf74a3" [ 946.062090] env[61857]: _type = "Task" [ 946.062090] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.072611] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52abe30b-86c4-0a36-60fd-47823fcf74a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.266187] env[61857]: DEBUG nova.compute.manager [req-41111e9f-7ccf-44d7-80a5-98bdebc5ba62 req-be45bc67-333d-4887-8372-d0de43596ab8 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Received event network-changed-d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 946.266187] env[61857]: DEBUG nova.compute.manager [req-41111e9f-7ccf-44d7-80a5-98bdebc5ba62 req-be45bc67-333d-4887-8372-d0de43596ab8 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Refreshing instance network info cache due to event network-changed-d9f80f94-9be0-4d62-863f-7562c6c9e63e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 946.266187] env[61857]: DEBUG oslo_concurrency.lockutils [req-41111e9f-7ccf-44d7-80a5-98bdebc5ba62 req-be45bc67-333d-4887-8372-d0de43596ab8 service nova] Acquiring lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.266187] env[61857]: DEBUG oslo_concurrency.lockutils [req-41111e9f-7ccf-44d7-80a5-98bdebc5ba62 req-be45bc67-333d-4887-8372-d0de43596ab8 service nova] Acquired lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.266187] env[61857]: DEBUG nova.network.neutron [req-41111e9f-7ccf-44d7-80a5-98bdebc5ba62 req-be45bc67-333d-4887-8372-d0de43596ab8 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Refreshing network info cache for port d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 946.434784] env[61857]: DEBUG nova.network.neutron [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Updated VIF entry in instance network info cache for port a5b81295-b253-46a2-b1a7-78d459f101de. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 946.435937] env[61857]: DEBUG nova.network.neutron [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Updating instance_info_cache with network_info: [{"id": "a5b81295-b253-46a2-b1a7-78d459f101de", "address": "fa:16:3e:59:26:5d", "network": {"id": "c3318ddc-c0fb-491b-badf-007720f0c9e1", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1428298316-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8dc7382605d14c748e938080baf494ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5b81295-b2", "ovs_interfaceid": "a5b81295-b253-46a2-b1a7-78d459f101de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.457829] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c3c9df36-d523-4a0f-9484-4446ad8c4138 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 946.520896] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951308, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.042211} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.521238] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] ce5f713d-8b63-4b95-a5ad-feddb95f921d/ce5f713d-8b63-4b95-a5ad-feddb95f921d.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 946.521501] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 946.521767] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d03d850d-b574-4912-950c-f54eb92823c2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.527392] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 946.527392] env[61857]: value = "task-951313" [ 946.527392] env[61857]: _type = "Task" [ 946.527392] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.535874] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951313, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.548706] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951311, 'name': ReconfigVM_Task, 'duration_secs': 0.57992} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.548996] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Reconfigured VM instance instance-00000055 to attach disk [datastore1] d85f8a2d-6053-45ec-81bb-de9fe02af3fe/d85f8a2d-6053-45ec-81bb-de9fe02af3fe.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.549641] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-117826bb-3133-4837-8c97-ea7f76d0f506 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.554721] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 946.554721] env[61857]: value = "task-951314" [ 946.554721] env[61857]: _type = "Task" [ 946.554721] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.564293] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951314, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.572154] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52abe30b-86c4-0a36-60fd-47823fcf74a3, 'name': SearchDatastore_Task, 'duration_secs': 0.014061} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.572460] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.572743] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.573014] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.573190] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.573406] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.573696] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69dab616-7839-497c-9ca9-87fd3846ac05 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.581934] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.582199] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 946.583013] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbcac965-c478-44ed-b23f-e822b99806c8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.588383] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for the task: (returnval){ [ 946.588383] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52296ce7-c445-f84e-c578-f0615b4c0f13" [ 946.588383] env[61857]: _type = "Task" [ 946.588383] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.596170] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52296ce7-c445-f84e-c578-f0615b4c0f13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.834252] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77385691-81ed-4f71-8394-b47429c8036c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.843607] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Doing hard reboot of VM {{(pid=61857) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 946.844185] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-217a3d8b-3b80-4e38-9bf8-ce0b7bfd0436 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.849683] env[61857]: DEBUG oslo_vmware.api [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 946.849683] env[61857]: value = "task-951315" [ 946.849683] env[61857]: _type = "Task" [ 946.849683] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.859021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.859021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.859021] env[61857]: DEBUG nova.network.neutron [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 946.860734] env[61857]: DEBUG oslo_vmware.api [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951315, 'name': ResetVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.933565] env[61857]: DEBUG nova.compute.manager [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 946.938345] env[61857]: DEBUG oslo_concurrency.lockutils [req-3d153697-fb53-4135-a1ac-749f0eef842c req-200bc78e-e666-4e18-a7fa-629892f511af service nova] Releasing lock "refresh_cache-c821ec43-9462-4276-bdc9-d54e64c0a3cf" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.959398] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 946.959658] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 946.959823] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.960030] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 946.960195] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.960435] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 946.960655] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 946.960823] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 946.960998] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 946.961250] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 946.961465] env[61857]: DEBUG nova.virt.hardware [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.962166] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance fd43f45b-0f54-46c6-94f9-ce28d3b15bb1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 946.962409] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 946.962639] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 946.965294] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa4529a-8027-4705-8968-2788776472de {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.975711] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb4294e-98f3-42d9-9a01-72f35d7a0f5e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.004656] env[61857]: DEBUG nova.network.neutron [req-41111e9f-7ccf-44d7-80a5-98bdebc5ba62 req-be45bc67-333d-4887-8372-d0de43596ab8 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updated VIF entry in instance network info cache for port d9f80f94-9be0-4d62-863f-7562c6c9e63e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 947.005127] env[61857]: DEBUG nova.network.neutron [req-41111e9f-7ccf-44d7-80a5-98bdebc5ba62 req-be45bc67-333d-4887-8372-d0de43596ab8 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updating instance_info_cache with network_info: [{"id": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "address": "fa:16:3e:94:19:6b", "network": {"id": "f9599202-4e91-4bab-bef6-17a31f285fe1", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2017759293-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3eca39f9d1f244a6b73b33326f4eea7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9f80f94-9b", "ovs_interfaceid": "d9f80f94-9be0-4d62-863f-7562c6c9e63e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.042037] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951313, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066712} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.042037] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 947.042037] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8392d89-8e80-41ae-b97b-a50768cdcb8e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.069029] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] ce5f713d-8b63-4b95-a5ad-feddb95f921d/ce5f713d-8b63-4b95-a5ad-feddb95f921d.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.073776] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50678443-4c3b-46e2-b49f-720c547f0cde {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.096550] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951314, 'name': Rename_Task, 'duration_secs': 0.378666} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.097077] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 947.097077] env[61857]: value = "task-951316" [ 947.097077] env[61857]: _type = "Task" [ 947.097077] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.099839] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 947.100378] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d6e9d68-d92a-427d-9e71-363e5698c853 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.110285] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52296ce7-c445-f84e-c578-f0615b4c0f13, 'name': SearchDatastore_Task, 'duration_secs': 0.024997} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.113675] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88021418-cc00-4279-851e-d6139f70d98f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.120021] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 947.120021] env[61857]: value = "task-951317" [ 947.120021] env[61857]: _type = "Task" [ 947.120021] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.120021] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951316, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.123889] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for the task: (returnval){ [ 947.123889] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52f40cc5-2e00-07d4-c0cf-f18f06ae741d" [ 947.123889] env[61857]: _type = "Task" [ 947.123889] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.131650] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951317, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.136181] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f40cc5-2e00-07d4-c0cf-f18f06ae741d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.238762] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a3cf7e-c697-45ae-b8d9-15aa8b6b1e6e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.246697] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c71af53-b75d-4cca-9020-60475b47ddc9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.277236] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0af90a6-4ccd-4253-8703-82f81dbc511b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.285061] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1e0b17-3ecb-4d33-9bcb-2b9f446469e4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.299395] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.360814] env[61857]: DEBUG oslo_vmware.api [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951315, 'name': ResetVM_Task, 'duration_secs': 0.092622} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.361182] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Did hard reboot of VM {{(pid=61857) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 947.361437] env[61857]: DEBUG nova.compute.manager [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.364095] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd71660-8384-47d8-8d26-d8af5f7f4aea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.510937] env[61857]: DEBUG oslo_concurrency.lockutils [req-41111e9f-7ccf-44d7-80a5-98bdebc5ba62 req-be45bc67-333d-4887-8372-d0de43596ab8 service nova] Releasing lock "refresh_cache-d9d05629-79ef-45c6-ac54-b3bc18d306f4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.613067] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951316, 'name': ReconfigVM_Task, 'duration_secs': 0.329861} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.613067] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Reconfigured VM instance instance-00000057 to attach disk [datastore1] ce5f713d-8b63-4b95-a5ad-feddb95f921d/ce5f713d-8b63-4b95-a5ad-feddb95f921d.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.613067] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a3e3ed0-fb08-4572-8702-66087a0540ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.619383] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 947.619383] env[61857]: value = "task-951318" [ 947.619383] env[61857]: _type = "Task" [ 947.619383] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.633553] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951318, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.636910] env[61857]: DEBUG oslo_vmware.api [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951317, 'name': PowerOnVM_Task, 'duration_secs': 0.457225} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.637338] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 947.637684] env[61857]: DEBUG nova.compute.manager [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.641775] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f2f9a4-87e1-4468-8fc4-995ba92410db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.644713] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f40cc5-2e00-07d4-c0cf-f18f06ae741d, 'name': SearchDatastore_Task, 'duration_secs': 0.026391} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.645135] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.645545] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] c821ec43-9462-4276-bdc9-d54e64c0a3cf/c821ec43-9462-4276-bdc9-d54e64c0a3cf.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 947.646273] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70e68fe4-cadf-4217-9e8f-b0ad871979e1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.659290] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for the task: (returnval){ [ 947.659290] env[61857]: value = "task-951319" [ 947.659290] env[61857]: _type = "Task" [ 947.659290] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.668427] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951319, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.802916] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.876623] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2e87b6d4-bc5b-4b85-af63-130cdf11d96b tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.852s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.893594] env[61857]: DEBUG nova.compute.manager [req-9fe1ae63-2c6e-4f36-9659-c14fe380fbcb req-f4d3a09f-abee-431d-9e84-006169259ed4 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received event network-vif-plugged-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.893824] env[61857]: DEBUG oslo_concurrency.lockutils [req-9fe1ae63-2c6e-4f36-9659-c14fe380fbcb req-f4d3a09f-abee-431d-9e84-006169259ed4 service nova] Acquiring lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.894048] env[61857]: DEBUG oslo_concurrency.lockutils [req-9fe1ae63-2c6e-4f36-9659-c14fe380fbcb req-f4d3a09f-abee-431d-9e84-006169259ed4 service nova] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.894225] env[61857]: DEBUG oslo_concurrency.lockutils [req-9fe1ae63-2c6e-4f36-9659-c14fe380fbcb req-f4d3a09f-abee-431d-9e84-006169259ed4 service nova] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.894397] env[61857]: DEBUG nova.compute.manager [req-9fe1ae63-2c6e-4f36-9659-c14fe380fbcb req-f4d3a09f-abee-431d-9e84-006169259ed4 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] No waiting events found dispatching network-vif-plugged-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.894563] env[61857]: WARNING nova.compute.manager [req-9fe1ae63-2c6e-4f36-9659-c14fe380fbcb req-f4d3a09f-abee-431d-9e84-006169259ed4 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received unexpected event network-vif-plugged-507983f7-c4c5-4605-887e-c3a1936e8f3a for instance with vm_state building and task_state spawning. [ 947.916229] env[61857]: DEBUG nova.network.neutron [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance_info_cache with network_info: [{"id": "e4c39a04-24d7-48d1-a089-2304778bd059", "address": "fa:16:3e:45:7a:cc", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4c39a04-24", "ovs_interfaceid": "e4c39a04-24d7-48d1-a089-2304778bd059", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.930370] env[61857]: DEBUG nova.network.neutron [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Successfully updated port: 507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.961927] env[61857]: DEBUG nova.compute.manager [req-c9b00a69-8fcf-476c-aadc-656afd283d75 req-636277d4-80a7-416f-b992-7050b7282d9b service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received event network-changed-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.961927] env[61857]: DEBUG nova.compute.manager [req-c9b00a69-8fcf-476c-aadc-656afd283d75 req-636277d4-80a7-416f-b992-7050b7282d9b service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Refreshing instance network info cache due to event network-changed-507983f7-c4c5-4605-887e-c3a1936e8f3a. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 947.961927] env[61857]: DEBUG oslo_concurrency.lockutils [req-c9b00a69-8fcf-476c-aadc-656afd283d75 req-636277d4-80a7-416f-b992-7050b7282d9b service nova] Acquiring lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.961927] env[61857]: DEBUG oslo_concurrency.lockutils [req-c9b00a69-8fcf-476c-aadc-656afd283d75 req-636277d4-80a7-416f-b992-7050b7282d9b service nova] Acquired lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.961927] env[61857]: DEBUG nova.network.neutron [req-c9b00a69-8fcf-476c-aadc-656afd283d75 req-636277d4-80a7-416f-b992-7050b7282d9b service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Refreshing network info cache for port 507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 948.129940] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951318, 'name': Rename_Task, 'duration_secs': 0.162866} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.130291] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 948.130565] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e3b6bae-d159-463b-b940-b2f31cb6e921 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.138043] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 948.138043] env[61857]: value = "task-951320" [ 948.138043] env[61857]: _type = "Task" [ 948.138043] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.147427] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951320, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.164798] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.170113] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951319, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.311111] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 948.311388] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.407s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.311717] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.154s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.311928] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.314248] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.488s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.314465] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.317153] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.868s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.319724] env[61857]: INFO nova.compute.claims [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 948.349303] env[61857]: INFO nova.scheduler.client.report [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted allocations for instance 992bcab3-7396-495f-96a1-e99ecd13c961 [ 948.418810] env[61857]: DEBUG oslo_concurrency.lockutils [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.433104] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.469199] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Volume attach. Driver type: vmdk {{(pid=61857) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 948.469553] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214170', 'volume_id': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'name': 'volume-e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0de8e67e-24b9-48d0-ac90-b5ec5b93e243', 'attached_at': '', 'detached_at': '', 'volume_id': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'serial': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 948.470732] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58af9218-2dc5-4da8-a55e-8697f04fa669 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.490168] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4055111-196c-466f-8d58-75d6e4f26d33 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.515908] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] volume-e2d3aded-f562-44fa-b5fc-9f83f20bc32d/volume-e2d3aded-f562-44fa-b5fc-9f83f20bc32d.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.516927] env[61857]: DEBUG nova.network.neutron [req-c9b00a69-8fcf-476c-aadc-656afd283d75 req-636277d4-80a7-416f-b992-7050b7282d9b service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 948.518701] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02b004ff-7c91-4814-8018-9e48377baea7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.538552] env[61857]: DEBUG oslo_vmware.api [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 948.538552] env[61857]: value = "task-951321" [ 948.538552] env[61857]: _type = "Task" [ 948.538552] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.547914] env[61857]: DEBUG oslo_vmware.api [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951321, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.635041] env[61857]: DEBUG nova.network.neutron [req-c9b00a69-8fcf-476c-aadc-656afd283d75 req-636277d4-80a7-416f-b992-7050b7282d9b service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.657284] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951320, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.671024] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951319, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.827877} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.671311] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] c821ec43-9462-4276-bdc9-d54e64c0a3cf/c821ec43-9462-4276-bdc9-d54e64c0a3cf.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 948.671642] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.671989] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0db7e0a8-00d8-48cf-be6d-0f569a44159c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.679950] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for the task: (returnval){ [ 948.679950] env[61857]: value = "task-951322" [ 948.679950] env[61857]: _type = "Task" [ 948.679950] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.690184] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951322, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.830731] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7fef1282-b303-4c69-b692-6116f610d2bf tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 35.900s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.831822] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 13.591s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.831952] env[61857]: INFO nova.compute.manager [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Unshelving [ 948.861344] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5544011-15e2-415e-bd47-d9523a2f6337 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "992bcab3-7396-495f-96a1-e99ecd13c961" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.660s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.943083] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf60f43-6210-43ec-814e-f1c1511a276a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.963582] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28db134c-f657-4c8f-9c24-657b7630a861 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.970968] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance '18a632be-322f-48a9-becf-51ff3b735ad4' progress to 83 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 949.048267] env[61857]: DEBUG oslo_vmware.api [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951321, 'name': ReconfigVM_Task, 'duration_secs': 0.368906} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.048571] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Reconfigured VM instance instance-00000051 to attach disk [datastore1] volume-e2d3aded-f562-44fa-b5fc-9f83f20bc32d/volume-e2d3aded-f562-44fa-b5fc-9f83f20bc32d.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.053342] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49c9fff8-db59-4712-a444-e79d75816f96 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.069046] env[61857]: DEBUG oslo_vmware.api [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 949.069046] env[61857]: value = "task-951323" [ 949.069046] env[61857]: _type = "Task" [ 949.069046] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.078214] env[61857]: DEBUG oslo_vmware.api [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951323, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.137596] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquiring lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.137916] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.138235] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquiring lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.138461] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.138647] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.140952] env[61857]: DEBUG oslo_concurrency.lockutils [req-c9b00a69-8fcf-476c-aadc-656afd283d75 req-636277d4-80a7-416f-b992-7050b7282d9b service nova] Releasing lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.144343] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.144566] env[61857]: DEBUG nova.network.neutron [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 949.147059] env[61857]: INFO nova.compute.manager [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Terminating instance [ 949.149255] env[61857]: DEBUG nova.compute.manager [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.149618] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.150630] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0b3b13-d5b6-4c8a-be57-9c49c89349c3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.157507] env[61857]: DEBUG oslo_vmware.api [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951320, 'name': PowerOnVM_Task, 'duration_secs': 0.732473} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.158074] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 949.158300] env[61857]: INFO nova.compute.manager [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Took 8.93 seconds to spawn the instance on the hypervisor. [ 949.158487] env[61857]: DEBUG nova.compute.manager [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.159484] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ffd153f-fd7f-4733-98be-e7059bda9605 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.164079] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.164649] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-867ac303-e181-4ba4-94b3-6a74cf09870e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.178536] env[61857]: DEBUG oslo_vmware.api [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 949.178536] env[61857]: value = "task-951324" [ 949.178536] env[61857]: _type = "Task" [ 949.178536] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.192130] env[61857]: DEBUG oslo_vmware.api [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951324, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.195079] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951322, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074404} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.195355] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.196169] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b571db67-61d4-4626-92a8-45643c99b78d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.220026] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] c821ec43-9462-4276-bdc9-d54e64c0a3cf/c821ec43-9462-4276-bdc9-d54e64c0a3cf.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.220265] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6daec4c4-db27-4fc5-8068-3a2a893beac3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.240668] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for the task: (returnval){ [ 949.240668] env[61857]: value = "task-951325" [ 949.240668] env[61857]: _type = "Task" [ 949.240668] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.248274] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951325, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.479657] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 949.479657] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94d2189a-c38f-4efe-aeaa-e3407c6f740b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.488319] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 949.488319] env[61857]: value = "task-951326" [ 949.488319] env[61857]: _type = "Task" [ 949.488319] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.496428] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951326, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.541961] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a284f6c-6751-4489-bb01-d0b68b9e9b36 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.548918] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "ce67a15a-8604-4523-a8fa-a34ccf9914da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.549180] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "ce67a15a-8604-4523-a8fa-a34ccf9914da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.553405] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcc59f8-2669-4ece-a071-30cb6abf45c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.589682] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2735b7-7965-453b-aaec-faa5bd158d4a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.599302] env[61857]: DEBUG oslo_vmware.api [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951323, 'name': ReconfigVM_Task, 'duration_secs': 0.141226} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.601654] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214170', 'volume_id': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'name': 'volume-e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0de8e67e-24b9-48d0-ac90-b5ec5b93e243', 'attached_at': '', 'detached_at': '', 'volume_id': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'serial': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 949.604409] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081860ca-d772-46c2-a705-d1012deb1b7c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.619454] env[61857]: DEBUG nova.compute.provider_tree [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.674783] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.675103] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.675378] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.675645] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.675860] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.681625] env[61857]: INFO nova.compute.manager [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Terminating instance [ 949.683486] env[61857]: DEBUG nova.network.neutron [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 949.686811] env[61857]: DEBUG nova.compute.manager [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.687023] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.690849] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb40957-65ca-4150-a22d-393dbce51231 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.695502] env[61857]: INFO nova.compute.manager [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Took 28.50 seconds to build instance. [ 949.702028] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.704473] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3740b30-bae3-49ce-af21-45cbf7eaa2ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.706055] env[61857]: DEBUG oslo_vmware.api [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951324, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.712929] env[61857]: DEBUG oslo_vmware.api [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 949.712929] env[61857]: value = "task-951327" [ 949.712929] env[61857]: _type = "Task" [ 949.712929] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.721554] env[61857]: DEBUG oslo_vmware.api [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951327, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.750115] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.831251] env[61857]: DEBUG nova.network.neutron [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating instance_info_cache with network_info: [{"id": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "address": "fa:16:3e:52:2f:3f", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507983f7-c4", "ovs_interfaceid": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.854475] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.000382] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951326, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.054121] env[61857]: DEBUG nova.compute.manager [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 950.122926] env[61857]: DEBUG nova.scheduler.client.report [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.197939] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7e574c3b-76ac-4776-a47f-2146ae81b717 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.015s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.198306] env[61857]: DEBUG oslo_vmware.api [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951324, 'name': PowerOffVM_Task, 'duration_secs': 0.526846} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.198650] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.198914] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.199312] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77d55c97-ba81-4686-9d38-7e81edfb50c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.226937] env[61857]: DEBUG oslo_vmware.api [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951327, 'name': PowerOffVM_Task, 'duration_secs': 0.319606} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.227791] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.227791] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.227791] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e776aec1-287d-47ba-8211-9ba548d0c6bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.253341] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951325, 'name': ReconfigVM_Task, 'duration_secs': 0.781345} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.253485] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Reconfigured VM instance instance-00000058 to attach disk [datastore1] c821ec43-9462-4276-bdc9-d54e64c0a3cf/c821ec43-9462-4276-bdc9-d54e64c0a3cf.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.254244] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-565e053e-a1e7-4f90-85da-154359116d61 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.259672] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for the task: (returnval){ [ 950.259672] env[61857]: value = "task-951330" [ 950.259672] env[61857]: _type = "Task" [ 950.259672] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.267748] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951330, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.270936] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.271152] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.271368] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Deleting the datastore file [datastore2] 825f5d87-76b1-4e4e-9b3d-4be419c7d323 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.271626] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-deb391d5-0fee-4505-9655-137a91a416be {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.278385] env[61857]: DEBUG oslo_vmware.api [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for the task: (returnval){ [ 950.278385] env[61857]: value = "task-951331" [ 950.278385] env[61857]: _type = "Task" [ 950.278385] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.286580] env[61857]: DEBUG oslo_vmware.api [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951331, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.301152] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.301444] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.301671] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleting the datastore file [datastore1] d85f8a2d-6053-45ec-81bb-de9fe02af3fe {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.301984] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-732b5582-b388-43d2-92df-9fa8536d03d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.308642] env[61857]: DEBUG oslo_vmware.api [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 950.308642] env[61857]: value = "task-951332" [ 950.308642] env[61857]: _type = "Task" [ 950.308642] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.316852] env[61857]: DEBUG oslo_vmware.api [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951332, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.334589] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.334957] env[61857]: DEBUG nova.compute.manager [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Instance network_info: |[{"id": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "address": "fa:16:3e:52:2f:3f", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507983f7-c4", "ovs_interfaceid": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 950.335447] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:2f:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '507983f7-c4c5-4605-887e-c3a1936e8f3a', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 950.344605] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating folder: Project (35ae3038a282412e963845e221e87f67). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 950.344946] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2bb14134-be20-4d28-85b3-8b7bb7154bc0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.354982] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Created folder: Project (35ae3038a282412e963845e221e87f67) in parent group-v214027. [ 950.355226] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating folder: Instances. Parent ref: group-v214174. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 950.355494] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0fdd19f2-348f-4510-a26c-162af45be6f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.364430] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Created folder: Instances in parent group-v214174. [ 950.364696] env[61857]: DEBUG oslo.service.loopingcall [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.364917] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 950.365151] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e963252d-3973-4806-9934-7257c3babe1f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.384404] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 950.384404] env[61857]: value = "task-951335" [ 950.384404] env[61857]: _type = "Task" [ 950.384404] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.394071] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951335, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.498579] env[61857]: DEBUG oslo_vmware.api [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951326, 'name': PowerOnVM_Task, 'duration_secs': 0.616285} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.498857] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 950.499063] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-015db394-2636-4dc6-8653-26d46a0d6267 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance '18a632be-322f-48a9-becf-51ff3b735ad4' progress to 100 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 950.576751] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.627744] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.627744] env[61857]: DEBUG nova.compute.manager [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 950.629901] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.072s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.631437] env[61857]: INFO nova.compute.claims [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.642543] env[61857]: DEBUG nova.objects.instance [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lazy-loading 'flavor' on Instance uuid 0de8e67e-24b9-48d0-ac90-b5ec5b93e243 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.769247] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951330, 'name': Rename_Task, 'duration_secs': 0.156202} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.769558] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 950.769820] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd1493df-eddc-46b9-9f97-b55175dab747 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.775239] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for the task: (returnval){ [ 950.775239] env[61857]: value = "task-951336" [ 950.775239] env[61857]: _type = "Task" [ 950.775239] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.785164] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.789881] env[61857]: DEBUG oslo_vmware.api [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Task: {'id': task-951331, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168763} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.790143] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.790364] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.790553] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.790728] env[61857]: INFO nova.compute.manager [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Took 1.64 seconds to destroy the instance on the hypervisor. [ 950.791046] env[61857]: DEBUG oslo.service.loopingcall [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.791232] env[61857]: DEBUG nova.compute.manager [-] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.791333] env[61857]: DEBUG nova.network.neutron [-] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.824233] env[61857]: DEBUG oslo_vmware.api [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951332, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18032} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.826484] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.826688] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.826890] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.827095] env[61857]: INFO nova.compute.manager [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Took 1.14 seconds to destroy the instance on the hypervisor. [ 950.827362] env[61857]: DEBUG oslo.service.loopingcall [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.827574] env[61857]: DEBUG nova.compute.manager [-] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.827672] env[61857]: DEBUG nova.network.neutron [-] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.894883] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951335, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.135653] env[61857]: DEBUG nova.compute.utils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 951.139104] env[61857]: DEBUG nova.compute.manager [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 951.139310] env[61857]: DEBUG nova.network.neutron [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 951.145764] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8510259c-d2fd-440e-b37c-807735028500 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.292s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.247430] env[61857]: DEBUG nova.policy [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74f5499cbed541deaf1905605e10158f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62bd314daba846cd9ed3dc994dd1034b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.257969] env[61857]: DEBUG nova.compute.manager [req-b5fb7f99-387c-43fa-bf7e-4c6c9f6dc96e req-854de62b-87dd-4067-b7d7-e0da76cfe1a9 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Received event network-vif-deleted-0b47d366-f4ff-4148-a2b4-be60b9228074 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.257969] env[61857]: INFO nova.compute.manager [req-b5fb7f99-387c-43fa-bf7e-4c6c9f6dc96e req-854de62b-87dd-4067-b7d7-e0da76cfe1a9 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Neutron deleted interface 0b47d366-f4ff-4148-a2b4-be60b9228074; detaching it from the instance and deleting it from the info cache [ 951.257969] env[61857]: DEBUG nova.network.neutron [req-b5fb7f99-387c-43fa-bf7e-4c6c9f6dc96e req-854de62b-87dd-4067-b7d7-e0da76cfe1a9 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.287199] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951336, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.322679] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.323066] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.397564] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951335, 'name': CreateVM_Task, 'duration_secs': 0.577475} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.397947] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 951.398915] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.399400] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.399903] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 951.400672] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbcf6b04-2303-4782-87cc-ae4e8e7a32ee {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.405923] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 951.405923] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52cee40c-8870-c00a-58f6-ee18addadefd" [ 951.405923] env[61857]: _type = "Task" [ 951.405923] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.415493] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52cee40c-8870-c00a-58f6-ee18addadefd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.578995] env[61857]: DEBUG nova.network.neutron [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Successfully created port: 5bd8b89d-fe07-48dc-a773-c222f7296a19 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 951.606268] env[61857]: DEBUG nova.compute.manager [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.607239] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c8c8eb-6533-4261-b9d7-d2f8431ecaf8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.644879] env[61857]: DEBUG nova.compute.manager [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 951.734259] env[61857]: DEBUG nova.network.neutron [-] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.741504] env[61857]: DEBUG nova.network.neutron [-] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.762041] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5374364f-a80a-4add-9481-ac50c63297ee {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.776390] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774f62f7-9444-4e36-a346-9d386c8db3c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.799957] env[61857]: DEBUG oslo_vmware.api [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951336, 'name': PowerOnVM_Task, 'duration_secs': 0.721734} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.800281] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 951.800500] env[61857]: INFO nova.compute.manager [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Took 9.17 seconds to spawn the instance on the hypervisor. [ 951.800685] env[61857]: DEBUG nova.compute.manager [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.801493] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7983f15-a0bc-440f-8dce-8f251e8f4b2e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.816767] env[61857]: DEBUG nova.compute.manager [req-b5fb7f99-387c-43fa-bf7e-4c6c9f6dc96e req-854de62b-87dd-4067-b7d7-e0da76cfe1a9 service nova] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Detach interface failed, port_id=0b47d366-f4ff-4148-a2b4-be60b9228074, reason: Instance d85f8a2d-6053-45ec-81bb-de9fe02af3fe could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 951.825898] env[61857]: INFO nova.compute.manager [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Detaching volume e2d3aded-f562-44fa-b5fc-9f83f20bc32d [ 951.877241] env[61857]: INFO nova.virt.block_device [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Attempting to driver detach volume e2d3aded-f562-44fa-b5fc-9f83f20bc32d from mountpoint /dev/sdb [ 951.878045] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Volume detach. Driver type: vmdk {{(pid=61857) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 951.878045] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214170', 'volume_id': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'name': 'volume-e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0de8e67e-24b9-48d0-ac90-b5ec5b93e243', 'attached_at': '', 'detached_at': '', 'volume_id': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'serial': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 951.878616] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6210d986-85cd-4b34-9c99-5cd6ef71af73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.906018] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8010f6-6d75-46d3-a546-baa1813e9a07 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.919306] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cdd21e2-c46b-4d57-8b01-baaddf73b985 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.921828] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52cee40c-8870-c00a-58f6-ee18addadefd, 'name': SearchDatastore_Task, 'duration_secs': 0.011127} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.924323] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.924640] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.924891] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.925061] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.925253] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.926071] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65c43ed7-ca3f-4c8b-bd72-2fac674f75e3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.946607] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f5a683-d78b-4b29-9272-33a38bfd2584 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.961747] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] The volume has not been displaced from its original location: [datastore1] volume-e2d3aded-f562-44fa-b5fc-9f83f20bc32d/volume-e2d3aded-f562-44fa-b5fc-9f83f20bc32d.vmdk. No consolidation needed. {{(pid=61857) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 951.967123] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Reconfiguring VM instance instance-00000051 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 951.970660] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a4816b3-c280-4f0f-980d-b960a56c6225 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.983379] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.983544] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 951.984556] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05ea27ff-b33a-4a87-94ed-2b953951dbb7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.991859] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 951.991859] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5273fe41-31f7-c46e-7949-0ad74fcaf965" [ 951.991859] env[61857]: _type = "Task" [ 951.991859] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.998082] env[61857]: DEBUG oslo_vmware.api [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 951.998082] env[61857]: value = "task-951337" [ 951.998082] env[61857]: _type = "Task" [ 951.998082] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.004340] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5273fe41-31f7-c46e-7949-0ad74fcaf965, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.009418] env[61857]: DEBUG oslo_vmware.api [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951337, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.025384] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c7dbe0-ccef-4c54-8263-1b3f8897bc26 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.032809] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5485bf0-affc-47fe-bc3f-4ae221e204be {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.064177] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a76f7db-8f2b-4c68-b712-a86391de9621 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.071782] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62b19af-3564-4748-879f-d7bac5dd42ae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.087079] env[61857]: DEBUG nova.compute.provider_tree [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 952.120814] env[61857]: INFO nova.compute.manager [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] instance snapshotting [ 952.124449] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a008d57d-7c79-4fda-898e-a3a6b9067e1f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.142926] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd932c54-c0c7-4c89-8211-13ea8e5a4571 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.235615] env[61857]: INFO nova.compute.manager [-] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Took 1.41 seconds to deallocate network for instance. [ 952.245713] env[61857]: INFO nova.compute.manager [-] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Took 1.45 seconds to deallocate network for instance. [ 952.331842] env[61857]: INFO nova.compute.manager [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Took 30.18 seconds to build instance. [ 952.509481] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5273fe41-31f7-c46e-7949-0ad74fcaf965, 'name': SearchDatastore_Task, 'duration_secs': 0.011497} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.511242] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5071ba2-1f16-4697-95b9-3399db8ebf0e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.519064] env[61857]: DEBUG oslo_vmware.api [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951337, 'name': ReconfigVM_Task, 'duration_secs': 0.2323} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.519935] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Reconfigured VM instance instance-00000051 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 952.529918] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b5f6f90-0338-41af-8ead-dc895bf0b40b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.547534] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 952.547534] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521dd5b1-63dc-650e-1231-76160fabd8f6" [ 952.547534] env[61857]: _type = "Task" [ 952.547534] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.556849] env[61857]: DEBUG oslo_vmware.api [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 952.556849] env[61857]: value = "task-951338" [ 952.556849] env[61857]: _type = "Task" [ 952.556849] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.567385] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521dd5b1-63dc-650e-1231-76160fabd8f6, 'name': SearchDatastore_Task, 'duration_secs': 0.033004} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.572436] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.573147] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] a42e8c77-3cc3-4323-b1ee-2a95192910d5/a42e8c77-3cc3-4323-b1ee-2a95192910d5.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 952.573147] env[61857]: DEBUG oslo_vmware.api [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951338, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.573439] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-edc07298-32b5-48dc-824f-7c145ae9c61a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.581580] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 952.581580] env[61857]: value = "task-951339" [ 952.581580] env[61857]: _type = "Task" [ 952.581580] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.596738] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951339, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.614922] env[61857]: ERROR nova.scheduler.client.report [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [req-e573f932-6498-4a6e-8c1c-c7138da042ef] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2d5860d0-8574-4e55-9ced-20e66f0314c2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e573f932-6498-4a6e-8c1c-c7138da042ef"}]} [ 952.632291] env[61857]: DEBUG nova.scheduler.client.report [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Refreshing inventories for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 952.646402] env[61857]: DEBUG nova.scheduler.client.report [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Updating ProviderTree inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 952.646637] env[61857]: DEBUG nova.compute.provider_tree [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 952.656043] env[61857]: DEBUG nova.compute.manager [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 952.658906] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 952.660348] env[61857]: DEBUG nova.scheduler.client.report [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Refreshing aggregate associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, aggregates: None {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 952.662130] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ef7b27e5-0617-4612-9ee7-6d6987893ef4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.669841] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 952.669841] env[61857]: value = "task-951340" [ 952.669841] env[61857]: _type = "Task" [ 952.669841] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.682141] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951340, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.684963] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 952.684963] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 952.685137] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.685384] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 952.685574] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.686201] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 952.687306] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 952.687306] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 952.688281] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 952.688281] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 952.688281] env[61857]: DEBUG nova.virt.hardware [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.690286] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56464502-164b-4b0a-8006-3fc34b9bf03a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.695900] env[61857]: DEBUG nova.scheduler.client.report [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Refreshing trait associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 952.707039] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79894aa1-0f59-4820-aef8-800b172eea35 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.743773] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.750268] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.834021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9c413341-c7d5-4d75-9d26-78079f86bbfc tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.697s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.928903] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e71a18e1-dd15-4a76-83b0-ea2fdf265a81 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.938312] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae4ee62-4143-41aa-8b95-fa586d858693 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.971669] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9e5131-44fc-41a4-af79-58ae64e4188e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.980601] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562e10f9-f006-4708-92ca-456e7b6a987f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.996194] env[61857]: DEBUG nova.compute.provider_tree [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 953.069553] env[61857]: DEBUG oslo_vmware.api [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951338, 'name': ReconfigVM_Task, 'duration_secs': 0.141396} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.069907] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214170', 'volume_id': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'name': 'volume-e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0de8e67e-24b9-48d0-ac90-b5ec5b93e243', 'attached_at': '', 'detached_at': '', 'volume_id': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d', 'serial': 'e2d3aded-f562-44fa-b5fc-9f83f20bc32d'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 953.093785] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951339, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.182216] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951340, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.204959] env[61857]: DEBUG oslo_concurrency.lockutils [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "18a632be-322f-48a9-becf-51ff3b735ad4" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.204959] env[61857]: DEBUG oslo_concurrency.lockutils [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "18a632be-322f-48a9-becf-51ff3b735ad4" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.204959] env[61857]: DEBUG nova.compute.manager [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Going to confirm migration 2 {{(pid=61857) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 953.291667] env[61857]: DEBUG nova.network.neutron [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Successfully updated port: 5bd8b89d-fe07-48dc-a773-c222f7296a19 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.298475] env[61857]: DEBUG nova.compute.manager [req-bf84f232-7cd3-412a-84f8-b73e543ada3b req-dca4bb1a-2ca9-450e-85ce-7d45b514de7a service nova] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Received event network-vif-deleted-553edf3c-e40b-4c94-b91c-f81a54df6643 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.298570] env[61857]: DEBUG nova.compute.manager [req-bf84f232-7cd3-412a-84f8-b73e543ada3b req-dca4bb1a-2ca9-450e-85ce-7d45b514de7a service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Received event network-vif-plugged-5bd8b89d-fe07-48dc-a773-c222f7296a19 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.298886] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf84f232-7cd3-412a-84f8-b73e543ada3b req-dca4bb1a-2ca9-450e-85ce-7d45b514de7a service nova] Acquiring lock "c3c9df36-d523-4a0f-9484-4446ad8c4138-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.299219] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf84f232-7cd3-412a-84f8-b73e543ada3b req-dca4bb1a-2ca9-450e-85ce-7d45b514de7a service nova] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.299524] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf84f232-7cd3-412a-84f8-b73e543ada3b req-dca4bb1a-2ca9-450e-85ce-7d45b514de7a service nova] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.299752] env[61857]: DEBUG nova.compute.manager [req-bf84f232-7cd3-412a-84f8-b73e543ada3b req-dca4bb1a-2ca9-450e-85ce-7d45b514de7a service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] No waiting events found dispatching network-vif-plugged-5bd8b89d-fe07-48dc-a773-c222f7296a19 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 953.300177] env[61857]: WARNING nova.compute.manager [req-bf84f232-7cd3-412a-84f8-b73e543ada3b req-dca4bb1a-2ca9-450e-85ce-7d45b514de7a service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Received unexpected event network-vif-plugged-5bd8b89d-fe07-48dc-a773-c222f7296a19 for instance with vm_state building and task_state spawning. [ 953.528687] env[61857]: DEBUG nova.scheduler.client.report [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Updated inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with generation 111 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 953.529023] env[61857]: DEBUG nova.compute.provider_tree [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Updating resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 generation from 111 to 112 during operation: update_inventory {{(pid=61857) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 953.529303] env[61857]: DEBUG nova.compute.provider_tree [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 953.547580] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquiring lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.547837] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.548063] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquiring lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.548278] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.548496] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.550613] env[61857]: INFO nova.compute.manager [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Terminating instance [ 953.552520] env[61857]: DEBUG nova.compute.manager [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 953.552728] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 953.553583] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da79cc8-a52e-441f-a2d5-d1ad56eb8c9f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.563862] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 953.563862] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c24a9ed-0819-4c89-8bcc-9561ce48498c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.570544] env[61857]: DEBUG oslo_vmware.api [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for the task: (returnval){ [ 953.570544] env[61857]: value = "task-951341" [ 953.570544] env[61857]: _type = "Task" [ 953.570544] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.579404] env[61857]: DEBUG oslo_vmware.api [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951341, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.593922] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951339, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640581} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.593922] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] a42e8c77-3cc3-4323-b1ee-2a95192910d5/a42e8c77-3cc3-4323-b1ee-2a95192910d5.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 953.593922] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.593922] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0436de64-b7f5-4310-81db-93ddd3270d4f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.600916] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 953.600916] env[61857]: value = "task-951342" [ 953.600916] env[61857]: _type = "Task" [ 953.600916] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.609192] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.622384] env[61857]: DEBUG nova.objects.instance [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lazy-loading 'flavor' on Instance uuid 0de8e67e-24b9-48d0-ac90-b5ec5b93e243 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.679940] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951340, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.753679] env[61857]: DEBUG oslo_concurrency.lockutils [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.753915] env[61857]: DEBUG oslo_concurrency.lockutils [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.754129] env[61857]: DEBUG nova.network.neutron [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 953.754329] env[61857]: DEBUG nova.objects.instance [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lazy-loading 'info_cache' on Instance uuid 18a632be-322f-48a9-becf-51ff3b735ad4 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.796502] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.796691] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.796952] env[61857]: DEBUG nova.network.neutron [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 954.035176] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.405s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.035743] env[61857]: DEBUG nova.compute.manager [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 954.038689] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 5.875s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.038918] env[61857]: DEBUG nova.objects.instance [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61857) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 954.081521] env[61857]: DEBUG oslo_vmware.api [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951341, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.110358] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.199287} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.110640] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.111424] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f04bae-e4a0-4f88-9adc-dfd32d3a2b8c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.132496] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] a42e8c77-3cc3-4323-b1ee-2a95192910d5/a42e8c77-3cc3-4323-b1ee-2a95192910d5.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.134366] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31542c97-b0d7-48cc-bd2d-0acc559cf591 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.153768] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 954.153768] env[61857]: value = "task-951343" [ 954.153768] env[61857]: _type = "Task" [ 954.153768] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.161743] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951343, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.180753] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951340, 'name': CreateSnapshot_Task, 'duration_secs': 1.159813} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.181010] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 954.181754] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95238427-2992-4139-b41c-82dac84465bd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.346845] env[61857]: DEBUG nova.network.neutron [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 954.513571] env[61857]: DEBUG nova.network.neutron [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Updating instance_info_cache with network_info: [{"id": "5bd8b89d-fe07-48dc-a773-c222f7296a19", "address": "fa:16:3e:71:32:7d", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bd8b89d-fe", "ovs_interfaceid": "5bd8b89d-fe07-48dc-a773-c222f7296a19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.543080] env[61857]: DEBUG nova.compute.utils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 954.546890] env[61857]: DEBUG nova.compute.manager [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 954.547073] env[61857]: DEBUG nova.network.neutron [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 954.581669] env[61857]: DEBUG oslo_vmware.api [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951341, 'name': PowerOffVM_Task, 'duration_secs': 0.520572} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.581937] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 954.582130] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 954.582393] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3478be7a-471b-4fd5-833a-e05622c355c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.584963] env[61857]: DEBUG nova.policy [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73092abfac2446a4a479b990e5309722', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb89b3d5d9c24fdd8263be171f21d115', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 954.650982] env[61857]: DEBUG oslo_concurrency.lockutils [None req-959c54ec-9c0f-4c4d-ad38-26a2f648cc48 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.328s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.665157] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.699575] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 954.699892] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 954.700100] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 954.700527] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Deleting the datastore file [datastore1] c821ec43-9462-4276-bdc9-d54e64c0a3cf {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.701123] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3fc4c4d8-f1ee-4323-bd81-f896cd8d23be {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.704580] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-840e40c3-5d5c-420e-a717-2fd4e2b7d472 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.711939] env[61857]: DEBUG oslo_vmware.api [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for the task: (returnval){ [ 954.711939] env[61857]: value = "task-951346" [ 954.711939] env[61857]: _type = "Task" [ 954.711939] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.712853] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 954.712853] env[61857]: value = "task-951345" [ 954.712853] env[61857]: _type = "Task" [ 954.712853] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.723950] env[61857]: DEBUG oslo_vmware.api [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.726897] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951345, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.016321] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.016670] env[61857]: DEBUG nova.compute.manager [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Instance network_info: |[{"id": "5bd8b89d-fe07-48dc-a773-c222f7296a19", "address": "fa:16:3e:71:32:7d", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bd8b89d-fe", "ovs_interfaceid": "5bd8b89d-fe07-48dc-a773-c222f7296a19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 955.017117] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:32:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5bd8b89d-fe07-48dc-a773-c222f7296a19', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.024787] env[61857]: DEBUG oslo.service.loopingcall [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.026416] env[61857]: DEBUG nova.network.neutron [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Successfully created port: 9c8f3dfe-a913-4426-9e25-9a2c4786fd82 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 955.027748] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 955.030019] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bdef6376-0c9c-4b10-ae7b-059556e05220 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.048542] env[61857]: DEBUG nova.compute.manager [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 955.050997] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 955.050997] env[61857]: value = "task-951347" [ 955.050997] env[61857]: _type = "Task" [ 955.050997] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.051896] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1ca6a07f-10df-479a-9eae-6e3ac32b81d5 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.053072] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.199s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.053291] env[61857]: DEBUG nova.objects.instance [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lazy-loading 'pci_requests' on Instance uuid 781a2790-e317-46fe-9be8-40c9e5f1f771 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.062582] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951347, 'name': CreateVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.165847] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951343, 'name': ReconfigVM_Task, 'duration_secs': 0.749182} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.166230] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Reconfigured VM instance instance-00000059 to attach disk [datastore2] a42e8c77-3cc3-4323-b1ee-2a95192910d5/a42e8c77-3cc3-4323-b1ee-2a95192910d5.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.166789] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dec84402-6044-4610-918c-a2140b58261d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.174482] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 955.174482] env[61857]: value = "task-951348" [ 955.174482] env[61857]: _type = "Task" [ 955.174482] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.186636] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951348, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.194033] env[61857]: DEBUG nova.network.neutron [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance_info_cache with network_info: [{"id": "e4c39a04-24d7-48d1-a089-2304778bd059", "address": "fa:16:3e:45:7a:cc", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4c39a04-24", "ovs_interfaceid": "e4c39a04-24d7-48d1-a089-2304778bd059", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.225154] env[61857]: DEBUG oslo_vmware.api [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Task: {'id': task-951346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147865} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.228497] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.228723] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 955.228924] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 955.229130] env[61857]: INFO nova.compute.manager [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Took 1.68 seconds to destroy the instance on the hypervisor. [ 955.229407] env[61857]: DEBUG oslo.service.loopingcall [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.230126] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951345, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.230126] env[61857]: DEBUG nova.compute.manager [-] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 955.230126] env[61857]: DEBUG nova.network.neutron [-] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 955.322997] env[61857]: DEBUG nova.compute.manager [req-3dc95397-2ee5-4d36-ba69-23aaddb6fc8a req-860d6d56-5e80-4cf2-b908-9278b1ad5388 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Received event network-changed-5bd8b89d-fe07-48dc-a773-c222f7296a19 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 955.323310] env[61857]: DEBUG nova.compute.manager [req-3dc95397-2ee5-4d36-ba69-23aaddb6fc8a req-860d6d56-5e80-4cf2-b908-9278b1ad5388 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Refreshing instance network info cache due to event network-changed-5bd8b89d-fe07-48dc-a773-c222f7296a19. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 955.323468] env[61857]: DEBUG oslo_concurrency.lockutils [req-3dc95397-2ee5-4d36-ba69-23aaddb6fc8a req-860d6d56-5e80-4cf2-b908-9278b1ad5388 service nova] Acquiring lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.323657] env[61857]: DEBUG oslo_concurrency.lockutils [req-3dc95397-2ee5-4d36-ba69-23aaddb6fc8a req-860d6d56-5e80-4cf2-b908-9278b1ad5388 service nova] Acquired lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.323776] env[61857]: DEBUG nova.network.neutron [req-3dc95397-2ee5-4d36-ba69-23aaddb6fc8a req-860d6d56-5e80-4cf2-b908-9278b1ad5388 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Refreshing network info cache for port 5bd8b89d-fe07-48dc-a773-c222f7296a19 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 955.565648] env[61857]: DEBUG nova.objects.instance [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lazy-loading 'numa_topology' on Instance uuid 781a2790-e317-46fe-9be8-40c9e5f1f771 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.579111] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951347, 'name': CreateVM_Task, 'duration_secs': 0.316909} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.579111] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 955.579546] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.579758] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.581771] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 955.582861] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d929273-2287-477e-bf5d-360a79c9cf8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.589606] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 955.589606] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52861227-83be-93d8-2480-62ec8a73293f" [ 955.589606] env[61857]: _type = "Task" [ 955.589606] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.607205] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52861227-83be-93d8-2480-62ec8a73293f, 'name': SearchDatastore_Task, 'duration_secs': 0.008839} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.610030] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.610030] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.612493] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.612493] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.612493] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.612786] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4de2353c-285c-4153-879b-b94cec833021 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.616237] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.618036] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.618036] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.618036] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.618036] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.623388] env[61857]: INFO nova.compute.manager [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Terminating instance [ 955.625595] env[61857]: DEBUG nova.compute.manager [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 955.625859] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 955.626958] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb3473ed-f04a-4229-9933-d6f46fe8a3f0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.631899] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.632148] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 955.633499] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd17d6c9-6a21-4e84-b3d3-bb1f46e080fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.639672] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 955.641730] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0403d43-6a8c-46b2-be05-87e202f12646 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.643726] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 955.643726] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52d27aca-306c-1335-393c-d5b011387f96" [ 955.643726] env[61857]: _type = "Task" [ 955.643726] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.649837] env[61857]: DEBUG oslo_vmware.api [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 955.649837] env[61857]: value = "task-951349" [ 955.649837] env[61857]: _type = "Task" [ 955.649837] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.660216] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52d27aca-306c-1335-393c-d5b011387f96, 'name': SearchDatastore_Task, 'duration_secs': 0.009269} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.661697] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f722573-f918-4502-87ec-7941d3e9d088 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.670098] env[61857]: DEBUG oslo_vmware.api [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951349, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.673833] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 955.673833] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5249cac2-5cf4-0b30-59d8-fb4116e13f2b" [ 955.673833] env[61857]: _type = "Task" [ 955.673833] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.688388] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5249cac2-5cf4-0b30-59d8-fb4116e13f2b, 'name': SearchDatastore_Task, 'duration_secs': 0.00881} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.692633] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.692877] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] c3c9df36-d523-4a0f-9484-4446ad8c4138/c3c9df36-d523-4a0f-9484-4446ad8c4138.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 955.693660] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951348, 'name': Rename_Task, 'duration_secs': 0.133621} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.695433] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f171d06c-8ced-4cc5-966c-8d72ab73a40e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.697985] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 955.698629] env[61857]: DEBUG oslo_concurrency.lockutils [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-18a632be-322f-48a9-becf-51ff3b735ad4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.698920] env[61857]: DEBUG nova.objects.instance [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lazy-loading 'migration_context' on Instance uuid 18a632be-322f-48a9-becf-51ff3b735ad4 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.700119] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2dfb5ab4-5ee2-44bf-95a0-b700898d3234 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.713862] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 955.713862] env[61857]: value = "task-951350" [ 955.713862] env[61857]: _type = "Task" [ 955.713862] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.714847] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 955.714847] env[61857]: value = "task-951351" [ 955.714847] env[61857]: _type = "Task" [ 955.714847] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.729789] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951351, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.734864] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951345, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.735154] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951350, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.046416] env[61857]: DEBUG nova.network.neutron [req-3dc95397-2ee5-4d36-ba69-23aaddb6fc8a req-860d6d56-5e80-4cf2-b908-9278b1ad5388 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Updated VIF entry in instance network info cache for port 5bd8b89d-fe07-48dc-a773-c222f7296a19. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 956.046806] env[61857]: DEBUG nova.network.neutron [req-3dc95397-2ee5-4d36-ba69-23aaddb6fc8a req-860d6d56-5e80-4cf2-b908-9278b1ad5388 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Updating instance_info_cache with network_info: [{"id": "5bd8b89d-fe07-48dc-a773-c222f7296a19", "address": "fa:16:3e:71:32:7d", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bd8b89d-fe", "ovs_interfaceid": "5bd8b89d-fe07-48dc-a773-c222f7296a19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.068445] env[61857]: DEBUG nova.compute.manager [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 956.070819] env[61857]: INFO nova.compute.claims [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.100877] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.101184] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.101355] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.101547] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.101700] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.101852] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.102076] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.102258] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.102451] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.102627] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.102814] env[61857]: DEBUG nova.virt.hardware [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.104109] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72802c75-92a2-4b2d-8293-803e02997fab {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.112676] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf732e0-7725-455a-9c5a-f2eb3a0ecab9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.160794] env[61857]: DEBUG oslo_vmware.api [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951349, 'name': PowerOffVM_Task, 'duration_secs': 0.165685} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.161195] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 956.161407] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 956.161682] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d34c637-25fc-4320-857c-a232b6b12f25 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.208847] env[61857]: DEBUG nova.objects.base [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Object Instance<18a632be-322f-48a9-becf-51ff3b735ad4> lazy-loaded attributes: info_cache,migration_context {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 956.209303] env[61857]: DEBUG nova.network.neutron [-] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.211416] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1c03f1-2a17-40cd-9da2-ca35140c9465 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.244040] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12b7b44c-dd66-45de-96ad-eb44dffc6233 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.247178] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951350, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.254769] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 956.255055] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 956.255247] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Deleting the datastore file [datastore1] 0de8e67e-24b9-48d0-ac90-b5ec5b93e243 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.255889] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951351, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.256152] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951345, 'name': CloneVM_Task, 'duration_secs': 1.437945} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.256936] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8962bf5c-a310-4b6d-b060-0fa0c2004a1f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.258715] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Created linked-clone VM from snapshot [ 956.260643] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be81723-bfa2-4301-9db8-bc84255b812d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.264376] env[61857]: DEBUG oslo_vmware.api [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 956.264376] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5255e12a-d8bb-fc14-2bb7-98411b9d30bf" [ 956.264376] env[61857]: _type = "Task" [ 956.264376] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.274162] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Uploading image a2c5dba8-94b6-4f0f-a3eb-8f079dbea813 {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 956.278296] env[61857]: DEBUG oslo_vmware.api [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for the task: (returnval){ [ 956.278296] env[61857]: value = "task-951353" [ 956.278296] env[61857]: _type = "Task" [ 956.278296] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.286254] env[61857]: DEBUG oslo_vmware.api [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5255e12a-d8bb-fc14-2bb7-98411b9d30bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.287312] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 956.287604] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b4919bf4-9022-4ee3-b9a1-dd7a358732f8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.293232] env[61857]: DEBUG oslo_vmware.api [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951353, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.299190] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 956.299190] env[61857]: value = "task-951354" [ 956.299190] env[61857]: _type = "Task" [ 956.299190] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.308167] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951354, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.550641] env[61857]: DEBUG oslo_concurrency.lockutils [req-3dc95397-2ee5-4d36-ba69-23aaddb6fc8a req-860d6d56-5e80-4cf2-b908-9278b1ad5388 service nova] Releasing lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.718137] env[61857]: INFO nova.compute.manager [-] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Took 1.49 seconds to deallocate network for instance. [ 956.738275] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951350, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.639339} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.739538] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] c3c9df36-d523-4a0f-9484-4446ad8c4138/c3c9df36-d523-4a0f-9484-4446ad8c4138.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 956.739538] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.739538] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d858423a-6f84-4c40-a565-b4168381f828 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.744368] env[61857]: DEBUG oslo_vmware.api [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951351, 'name': PowerOnVM_Task, 'duration_secs': 0.870829} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.745088] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 956.745450] env[61857]: INFO nova.compute.manager [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Took 9.81 seconds to spawn the instance on the hypervisor. [ 956.745659] env[61857]: DEBUG nova.compute.manager [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.746469] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715ad785-5d1d-44e9-a261-d55a4a8b863b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.753383] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 956.753383] env[61857]: value = "task-951355" [ 956.753383] env[61857]: _type = "Task" [ 956.753383] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.766964] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951355, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.778721] env[61857]: DEBUG oslo_vmware.api [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5255e12a-d8bb-fc14-2bb7-98411b9d30bf, 'name': SearchDatastore_Task, 'duration_secs': 0.022527} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.779653] env[61857]: DEBUG oslo_concurrency.lockutils [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.792785] env[61857]: DEBUG oslo_vmware.api [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Task: {'id': task-951353, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259392} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.792785] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.792785] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 956.793075] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 956.793121] env[61857]: INFO nova.compute.manager [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Took 1.17 seconds to destroy the instance on the hypervisor. [ 956.793495] env[61857]: DEBUG oslo.service.loopingcall [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.793596] env[61857]: DEBUG nova.compute.manager [-] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.793638] env[61857]: DEBUG nova.network.neutron [-] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 956.808413] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951354, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.855865] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46322896-34c3-46b1-bd2b-95674b9e6e4d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.863584] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3242e46-0921-4ba1-ab3d-964fa81879ff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.895446] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4871aeb2-43b4-4ff0-93d3-f3680a330ead {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.903134] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27222960-f868-4ea7-b616-259167157d28 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.917305] env[61857]: DEBUG nova.compute.provider_tree [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 957.078901] env[61857]: DEBUG nova.network.neutron [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Successfully updated port: 9c8f3dfe-a913-4426-9e25-9a2c4786fd82 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 957.229437] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.271161] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951355, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099635} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.273479] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.274089] env[61857]: INFO nova.compute.manager [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Took 33.54 seconds to build instance. [ 957.275546] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d65036f-b8c8-4762-b8fb-f37b851c2bdd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.308441] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] c3c9df36-d523-4a0f-9484-4446ad8c4138/c3c9df36-d523-4a0f-9484-4446ad8c4138.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.309934] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-402abdea-64e5-46ec-a55f-4b74358a915c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.334312] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951354, 'name': Destroy_Task, 'duration_secs': 0.946728} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.335806] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Destroyed the VM [ 957.336097] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 957.336457] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 957.336457] env[61857]: value = "task-951356" [ 957.336457] env[61857]: _type = "Task" [ 957.336457] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.336682] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1f18a3ef-204c-4274-bf60-45328802fb41 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.349223] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951356, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.350591] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 957.350591] env[61857]: value = "task-951357" [ 957.350591] env[61857]: _type = "Task" [ 957.350591] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.356691] env[61857]: DEBUG nova.compute.manager [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Received event network-vif-deleted-a5b81295-b253-46a2-b1a7-78d459f101de {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.356904] env[61857]: DEBUG nova.compute.manager [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Received event network-vif-plugged-9c8f3dfe-a913-4426-9e25-9a2c4786fd82 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.357124] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] Acquiring lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.357350] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] Lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.357534] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] Lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.357726] env[61857]: DEBUG nova.compute.manager [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] No waiting events found dispatching network-vif-plugged-9c8f3dfe-a913-4426-9e25-9a2c4786fd82 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 957.357907] env[61857]: WARNING nova.compute.manager [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Received unexpected event network-vif-plugged-9c8f3dfe-a913-4426-9e25-9a2c4786fd82 for instance with vm_state building and task_state spawning. [ 957.358096] env[61857]: DEBUG nova.compute.manager [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Received event network-changed-9c8f3dfe-a913-4426-9e25-9a2c4786fd82 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.358269] env[61857]: DEBUG nova.compute.manager [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Refreshing instance network info cache due to event network-changed-9c8f3dfe-a913-4426-9e25-9a2c4786fd82. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 957.358470] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] Acquiring lock "refresh_cache-fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.358675] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] Acquired lock "refresh_cache-fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.358810] env[61857]: DEBUG nova.network.neutron [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Refreshing network info cache for port 9c8f3dfe-a913-4426-9e25-9a2c4786fd82 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 957.365311] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951357, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.464417] env[61857]: DEBUG nova.scheduler.client.report [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Updated inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with generation 112 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 957.464417] env[61857]: DEBUG nova.compute.provider_tree [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Updating resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 generation from 112 to 113 during operation: update_inventory {{(pid=61857) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 957.464417] env[61857]: DEBUG nova.compute.provider_tree [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 957.581442] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquiring lock "refresh_cache-fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.780029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e62e91fc-bc33-4ab5-81e3-17ee3632b3f7 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.061s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.851608] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951356, 'name': ReconfigVM_Task, 'duration_secs': 0.310261} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.857380] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Reconfigured VM instance instance-0000005a to attach disk [datastore2] c3c9df36-d523-4a0f-9484-4446ad8c4138/c3c9df36-d523-4a0f-9484-4446ad8c4138.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.857978] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53b034c4-0c54-40b5-bfd7-f83eebefd4e7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.867473] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951357, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.867771] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 957.867771] env[61857]: value = "task-951358" [ 957.867771] env[61857]: _type = "Task" [ 957.867771] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.875725] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951358, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.963530] env[61857]: DEBUG nova.network.neutron [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 957.966076] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.913s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.968131] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.391s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.969787] env[61857]: INFO nova.compute.claims [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.075989] env[61857]: INFO nova.network.neutron [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updating port 5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 958.256527] env[61857]: DEBUG nova.network.neutron [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.361440] env[61857]: DEBUG oslo_vmware.api [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951357, 'name': RemoveSnapshot_Task, 'duration_secs': 0.678506} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.361753] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 958.377519] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951358, 'name': Rename_Task, 'duration_secs': 0.167778} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.377790] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 958.378045] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07fc1e9f-2b8c-4569-a9fb-4f96cba5f178 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.387021] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 958.387021] env[61857]: value = "task-951359" [ 958.387021] env[61857]: _type = "Task" [ 958.387021] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.396017] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951359, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.630174] env[61857]: DEBUG nova.network.neutron [-] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.759385] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f94bedc-871a-4acb-a760-4c78bf5b2021 req-b580b442-1f6e-4130-8a92-2b511ad4c462 service nova] Releasing lock "refresh_cache-fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.759835] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquired lock "refresh_cache-fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.759999] env[61857]: DEBUG nova.network.neutron [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 958.866861] env[61857]: WARNING nova.compute.manager [None req-f469f334-700e-4316-a545-f803e1ade330 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Image not found during snapshot: nova.exception.ImageNotFound: Image a2c5dba8-94b6-4f0f-a3eb-8f079dbea813 could not be found. [ 958.897288] env[61857]: DEBUG oslo_vmware.api [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951359, 'name': PowerOnVM_Task, 'duration_secs': 0.453345} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.897573] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 958.897779] env[61857]: INFO nova.compute.manager [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Took 6.24 seconds to spawn the instance on the hypervisor. [ 958.897969] env[61857]: DEBUG nova.compute.manager [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.899038] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa531f9-4d0c-4cd9-85da-22e868372890 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.132211] env[61857]: INFO nova.compute.manager [-] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Took 2.34 seconds to deallocate network for instance. [ 959.209971] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc47575-3a86-46ca-9b2a-4f3ac931edaf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.218881] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceff1d0b-1daf-4220-a344-e68dd0900732 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.249331] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee21c52-50af-4502-b9f6-dcabbe651ba5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.257783] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f397892-ce7d-4864-b78b-5e05885fbde0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.273591] env[61857]: DEBUG nova.compute.provider_tree [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.319857] env[61857]: DEBUG nova.network.neutron [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 959.418030] env[61857]: INFO nova.compute.manager [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Took 22.01 seconds to build instance. [ 959.493914] env[61857]: DEBUG nova.network.neutron [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Updating instance_info_cache with network_info: [{"id": "9c8f3dfe-a913-4426-9e25-9a2c4786fd82", "address": "fa:16:3e:65:b4:b3", "network": {"id": "fe5e4973-8006-40a0-957f-dbea64aa45c4", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1124042257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb89b3d5d9c24fdd8263be171f21d115", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "950a2f67-7668-4376-9d48-b38dca033c40", "external-id": "nsx-vlan-transportzone-549", "segmentation_id": 549, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c8f3dfe-a9", "ovs_interfaceid": "9c8f3dfe-a913-4426-9e25-9a2c4786fd82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.502486] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.502610] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.502925] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.503277] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.503474] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.505636] env[61857]: INFO nova.compute.manager [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Terminating instance [ 959.507666] env[61857]: DEBUG nova.compute.manager [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.507863] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 959.508709] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22674a63-a7af-436e-bf50-b7eb495e074b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.516846] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 959.517085] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2bb0114-bfa6-48e8-9bff-2b74a9ec5ba3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.522853] env[61857]: DEBUG oslo_vmware.api [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 959.522853] env[61857]: value = "task-951360" [ 959.522853] env[61857]: _type = "Task" [ 959.522853] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.531562] env[61857]: DEBUG oslo_vmware.api [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951360, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.595345] env[61857]: DEBUG nova.compute.manager [req-09c0b472-9f58-4b9a-8185-65f0fade2955 req-fb71b1cc-dec8-46d8-aaf8-b8e6c232d55a service nova] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Received event network-vif-deleted-639da037-e9cb-4624-994d-e7faab16e538 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 959.595345] env[61857]: DEBUG nova.compute.manager [req-09c0b472-9f58-4b9a-8185-65f0fade2955 req-fb71b1cc-dec8-46d8-aaf8-b8e6c232d55a service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received event network-changed-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 959.595449] env[61857]: DEBUG nova.compute.manager [req-09c0b472-9f58-4b9a-8185-65f0fade2955 req-fb71b1cc-dec8-46d8-aaf8-b8e6c232d55a service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Refreshing instance network info cache due to event network-changed-507983f7-c4c5-4605-887e-c3a1936e8f3a. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 959.595725] env[61857]: DEBUG oslo_concurrency.lockutils [req-09c0b472-9f58-4b9a-8185-65f0fade2955 req-fb71b1cc-dec8-46d8-aaf8-b8e6c232d55a service nova] Acquiring lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.595888] env[61857]: DEBUG oslo_concurrency.lockutils [req-09c0b472-9f58-4b9a-8185-65f0fade2955 req-fb71b1cc-dec8-46d8-aaf8-b8e6c232d55a service nova] Acquired lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.596070] env[61857]: DEBUG nova.network.neutron [req-09c0b472-9f58-4b9a-8185-65f0fade2955 req-fb71b1cc-dec8-46d8-aaf8-b8e6c232d55a service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Refreshing network info cache for port 507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 959.639476] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.776154] env[61857]: DEBUG nova.scheduler.client.report [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.920345] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b6619a01-1b2e-482b-899e-88561beaf755 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.523s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.997346] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Releasing lock "refresh_cache-fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.997346] env[61857]: DEBUG nova.compute.manager [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Instance network_info: |[{"id": "9c8f3dfe-a913-4426-9e25-9a2c4786fd82", "address": "fa:16:3e:65:b4:b3", "network": {"id": "fe5e4973-8006-40a0-957f-dbea64aa45c4", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1124042257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb89b3d5d9c24fdd8263be171f21d115", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "950a2f67-7668-4376-9d48-b38dca033c40", "external-id": "nsx-vlan-transportzone-549", "segmentation_id": 549, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c8f3dfe-a9", "ovs_interfaceid": "9c8f3dfe-a913-4426-9e25-9a2c4786fd82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 959.997684] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:b4:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '950a2f67-7668-4376-9d48-b38dca033c40', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c8f3dfe-a913-4426-9e25-9a2c4786fd82', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 960.005318] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Creating folder: Project (fb89b3d5d9c24fdd8263be171f21d115). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 960.005603] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2776111a-2f7d-4f88-b516-e6ec261f19f8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.016878] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Created folder: Project (fb89b3d5d9c24fdd8263be171f21d115) in parent group-v214027. [ 960.017089] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Creating folder: Instances. Parent ref: group-v214180. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 960.017330] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8a8fa7c9-7179-4621-87f7-ee5619e506b9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.027539] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Created folder: Instances in parent group-v214180. [ 960.027774] env[61857]: DEBUG oslo.service.loopingcall [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.028325] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 960.028619] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7acf4383-3684-4ed9-8a64-dd0ae08480a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.045766] env[61857]: DEBUG oslo_vmware.api [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951360, 'name': PowerOffVM_Task, 'duration_secs': 0.177668} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.046384] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 960.046563] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 960.046806] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4c26f2c-cf7e-447f-a2db-d7c9885bec9d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.051659] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 960.051659] env[61857]: value = "task-951363" [ 960.051659] env[61857]: _type = "Task" [ 960.051659] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.058705] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951363, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.111430] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 960.111690] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 960.111881] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleting the datastore file [datastore1] ce5f713d-8b63-4b95-a5ad-feddb95f921d {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.112181] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-833a25fa-6bad-4e4e-b50f-d4b2356b88ea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.118757] env[61857]: DEBUG oslo_vmware.api [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for the task: (returnval){ [ 960.118757] env[61857]: value = "task-951365" [ 960.118757] env[61857]: _type = "Task" [ 960.118757] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.126356] env[61857]: DEBUG oslo_vmware.api [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951365, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.282247] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.283090] env[61857]: DEBUG nova.compute.manager [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.287254] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.543s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.287673] env[61857]: DEBUG nova.objects.instance [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'resources' on Instance uuid d85f8a2d-6053-45ec-81bb-de9fe02af3fe {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.320635] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "c3c9df36-d523-4a0f-9484-4446ad8c4138" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.320961] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.321178] env[61857]: INFO nova.compute.manager [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Shelving [ 960.422046] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.422046] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.422046] env[61857]: DEBUG nova.network.neutron [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 960.498472] env[61857]: DEBUG nova.network.neutron [req-09c0b472-9f58-4b9a-8185-65f0fade2955 req-fb71b1cc-dec8-46d8-aaf8-b8e6c232d55a service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updated VIF entry in instance network info cache for port 507983f7-c4c5-4605-887e-c3a1936e8f3a. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 960.498908] env[61857]: DEBUG nova.network.neutron [req-09c0b472-9f58-4b9a-8185-65f0fade2955 req-fb71b1cc-dec8-46d8-aaf8-b8e6c232d55a service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating instance_info_cache with network_info: [{"id": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "address": "fa:16:3e:52:2f:3f", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507983f7-c4", "ovs_interfaceid": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.561635] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951363, 'name': CreateVM_Task, 'duration_secs': 0.385842} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.561799] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 960.562486] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.562713] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.562982] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 960.563245] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb7be863-a4e1-48a5-a0e4-a5a018a40cb0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.567605] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for the task: (returnval){ [ 960.567605] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5264e88f-8113-0a10-359b-1a52c090761b" [ 960.567605] env[61857]: _type = "Task" [ 960.567605] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.574697] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5264e88f-8113-0a10-359b-1a52c090761b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.628011] env[61857]: DEBUG oslo_vmware.api [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Task: {'id': task-951365, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245502} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.628291] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.628485] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 960.628662] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 960.628839] env[61857]: INFO nova.compute.manager [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 960.629100] env[61857]: DEBUG oslo.service.loopingcall [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.629310] env[61857]: DEBUG nova.compute.manager [-] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 960.629405] env[61857]: DEBUG nova.network.neutron [-] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 960.793073] env[61857]: DEBUG nova.compute.utils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.796893] env[61857]: DEBUG nova.compute.manager [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 960.796950] env[61857]: DEBUG nova.network.neutron [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 960.828684] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 960.829781] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4687ca74-5af7-4d93-8166-89c94809cef9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.836185] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 960.836185] env[61857]: value = "task-951366" [ 960.836185] env[61857]: _type = "Task" [ 960.836185] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.850839] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951366, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.859518] env[61857]: DEBUG nova.policy [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2ef7e845007475a8a19cf3f520c3f51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1acf476aabc4166bc8505a3442367c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 961.004879] env[61857]: DEBUG oslo_concurrency.lockutils [req-09c0b472-9f58-4b9a-8185-65f0fade2955 req-fb71b1cc-dec8-46d8-aaf8-b8e6c232d55a service nova] Releasing lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.081037] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5264e88f-8113-0a10-359b-1a52c090761b, 'name': SearchDatastore_Task, 'duration_secs': 0.044777} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.083531] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.083776] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 961.084025] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.084186] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.084374] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.084831] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-daddda5f-8c12-4bdb-9887-d91f47e96158 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.093278] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.093489] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 961.094226] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9de1c338-2cbc-4216-8ad6-3feea6c5a56c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.107410] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for the task: (returnval){ [ 961.107410] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52c576ab-60d5-5947-0f43-cc4e7eb5ef7d" [ 961.107410] env[61857]: _type = "Task" [ 961.107410] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.116823] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c576ab-60d5-5947-0f43-cc4e7eb5ef7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.120167] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c98c57-ec86-4341-882e-6b38a899c183 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.126793] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d1e9c6-6092-46f5-b444-ba4c534e8cb8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.156292] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec034372-ef73-4830-82a8-18930cf751fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.167568] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725085ab-db43-4bc8-b8a0-2d935166e8f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.186996] env[61857]: DEBUG nova.compute.provider_tree [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.300865] env[61857]: DEBUG nova.compute.manager [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 961.350640] env[61857]: DEBUG nova.network.neutron [-] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.352229] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951366, 'name': PowerOffVM_Task, 'duration_secs': 0.180757} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.352756] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 961.353617] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399613a5-c4cb-4185-9ffa-da61688b15d3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.357847] env[61857]: DEBUG nova.network.neutron [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updating instance_info_cache with network_info: [{"id": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "address": "fa:16:3e:dc:03:c9", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5da8191c-5f", "ovs_interfaceid": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.375074] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.377741] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f03f211-a055-47d4-a4df-fd6b5ab55a42 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.381138] env[61857]: DEBUG nova.network.neutron [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Successfully created port: 971c8b59-b659-4d32-9fba-17fdfffd9fe2 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.407935] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='22e8cb40d35516fd742474a7e28b07ca',container_format='bare',created_at=2024-09-18T21:44:36Z,direct_url=,disk_format='vmdk',id=efabb9a4-7907-4262-a138-86ad55c84b5b,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1138106088-shelved',owner='4725d163dec74a53940dc24ae88927f8',properties=ImageMetaProps,protected=,size=31660032,status='active',tags=,updated_at=2024-09-18T21:44:51Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 961.408245] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 961.408410] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.408601] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 961.408754] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.408907] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 961.409171] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 961.409356] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 961.409537] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 961.409707] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 961.409888] env[61857]: DEBUG nova.virt.hardware [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.410782] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22114194-d172-4309-88ba-3fd6870fe4a2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.419657] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee6762e-a1eb-4870-bf37-23e45351455f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.434115] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:03:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.443518] env[61857]: DEBUG oslo.service.loopingcall [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.443518] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 961.443518] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac8fbe78-7b0e-4633-b767-a924728db2fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.462605] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.462605] env[61857]: value = "task-951367" [ 961.462605] env[61857]: _type = "Task" [ 961.462605] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.470103] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951367, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.617995] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c576ab-60d5-5947-0f43-cc4e7eb5ef7d, 'name': SearchDatastore_Task, 'duration_secs': 0.010788} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.618833] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96d09c72-c674-4168-8975-34df6524a4f7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.624152] env[61857]: DEBUG nova.compute.manager [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Received event network-vif-plugged-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 961.624379] env[61857]: DEBUG oslo_concurrency.lockutils [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] Acquiring lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.624689] env[61857]: DEBUG oslo_concurrency.lockutils [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.624746] env[61857]: DEBUG oslo_concurrency.lockutils [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.626858] env[61857]: DEBUG nova.compute.manager [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] No waiting events found dispatching network-vif-plugged-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 961.626858] env[61857]: WARNING nova.compute.manager [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Received unexpected event network-vif-plugged-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 for instance with vm_state shelved_offloaded and task_state spawning. [ 961.626858] env[61857]: DEBUG nova.compute.manager [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Received event network-changed-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 961.626858] env[61857]: DEBUG nova.compute.manager [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Refreshing instance network info cache due to event network-changed-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 961.626858] env[61857]: DEBUG oslo_concurrency.lockutils [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] Acquiring lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.626858] env[61857]: DEBUG oslo_concurrency.lockutils [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] Acquired lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.626858] env[61857]: DEBUG nova.network.neutron [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Refreshing network info cache for port 5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 961.629518] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for the task: (returnval){ [ 961.629518] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52a83613-7da4-635d-119d-a9f99682a438" [ 961.629518] env[61857]: _type = "Task" [ 961.629518] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.640476] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a83613-7da4-635d-119d-a9f99682a438, 'name': SearchDatastore_Task, 'duration_secs': 0.010642} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.641262] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.641531] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] fd43f45b-0f54-46c6-94f9-ce28d3b15bb1/fd43f45b-0f54-46c6-94f9-ce28d3b15bb1.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 961.641792] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a427cc97-4275-4e8f-a12b-e660e1af3bfc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.648510] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for the task: (returnval){ [ 961.648510] env[61857]: value = "task-951368" [ 961.648510] env[61857]: _type = "Task" [ 961.648510] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.656852] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951368, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.690757] env[61857]: DEBUG nova.scheduler.client.report [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.854042] env[61857]: INFO nova.compute.manager [-] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Took 1.22 seconds to deallocate network for instance. [ 961.894048] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 961.894048] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c44006b8-498c-400e-9e9d-7d042405bf3b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.901958] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 961.901958] env[61857]: value = "task-951369" [ 961.901958] env[61857]: _type = "Task" [ 961.901958] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.914393] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951369, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.973905] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951367, 'name': CreateVM_Task, 'duration_secs': 0.400014} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.973905] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 961.974675] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.974845] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "[datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.975337] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 961.975627] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed8aa8bf-9d37-4c5d-bc90-d2b8bf8f58b7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.981639] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 961.981639] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52311773-c66d-f237-1705-97238e9a4020" [ 961.981639] env[61857]: _type = "Task" [ 961.981639] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.994925] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52311773-c66d-f237-1705-97238e9a4020, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.158680] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951368, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.195459] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.199465] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.449s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.199465] env[61857]: DEBUG nova.objects.instance [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lazy-loading 'resources' on Instance uuid 825f5d87-76b1-4e4e-9b3d-4be419c7d323 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.223330] env[61857]: INFO nova.scheduler.client.report [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleted allocations for instance d85f8a2d-6053-45ec-81bb-de9fe02af3fe [ 962.310300] env[61857]: DEBUG nova.compute.manager [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.346763] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.350129] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.350129] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.350129] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.350129] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.350129] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.350129] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.350129] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.350129] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.350129] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.350129] env[61857]: DEBUG nova.virt.hardware [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.350129] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7059348-5e15-4bea-968d-7c038c6720f6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.361197] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.362714] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b295f3-fc4c-415d-bb10-7ab973c6c346 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.413085] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951369, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.452607] env[61857]: DEBUG nova.network.neutron [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updated VIF entry in instance network info cache for port 5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.452991] env[61857]: DEBUG nova.network.neutron [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updating instance_info_cache with network_info: [{"id": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "address": "fa:16:3e:dc:03:c9", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5da8191c-5f", "ovs_interfaceid": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.491769] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "[datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.492044] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Processing image efabb9a4-7907-4262-a138-86ad55c84b5b {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.492292] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b/efabb9a4-7907-4262-a138-86ad55c84b5b.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.492447] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "[datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b/efabb9a4-7907-4262-a138-86ad55c84b5b.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.492691] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.492947] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fbd45f72-9448-484e-86b1-9f38d79cf805 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.503066] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.503196] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 962.503884] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a55e33f-35c2-40c3-9a27-799ec391ac96 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.509112] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 962.509112] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5202f494-95ea-da88-b4ae-63d447507e0f" [ 962.509112] env[61857]: _type = "Task" [ 962.509112] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.516695] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5202f494-95ea-da88-b4ae-63d447507e0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.660139] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951368, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522834} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.660425] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] fd43f45b-0f54-46c6-94f9-ce28d3b15bb1/fd43f45b-0f54-46c6-94f9-ce28d3b15bb1.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 962.660662] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 962.660932] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6cd71e4-5f12-46d8-a02c-1301ecc3837a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.667210] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for the task: (returnval){ [ 962.667210] env[61857]: value = "task-951370" [ 962.667210] env[61857]: _type = "Task" [ 962.667210] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.674394] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951370, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.731827] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ea9772e1-0937-47ac-8da8-49ed2b4cd11a tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "d85f8a2d-6053-45ec-81bb-de9fe02af3fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.056s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.911715] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951369, 'name': CreateSnapshot_Task, 'duration_secs': 0.528928} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.912964] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 962.913859] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b90340-9dca-41a7-afba-d0600632e9c5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.916877] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63efaf0a-3a86-4f99-8695-2996b2e352d1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.926428] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6d1c6c-9013-46df-8d83-93056c9bd4a7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.963614] env[61857]: DEBUG oslo_concurrency.lockutils [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] Releasing lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.964205] env[61857]: DEBUG nova.compute.manager [req-9adf641e-6793-438e-9035-4f7d9c2ffe36 req-f73acb67-349d-4149-a21a-7428e4a17c29 service nova] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Received event network-vif-deleted-d64de57d-2c7b-4898-8fc2-0ebf0bc06faf {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 962.965995] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c7d542-2c71-4dd7-bc4c-3374be4069b7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.974670] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549564af-61d5-4b69-9c36-a8c01c7ae128 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.989143] env[61857]: DEBUG nova.compute.provider_tree [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.019295] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Preparing fetch location {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 963.021944] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Fetch image to [datastore1] OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e/OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e.vmdk {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 963.021944] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Downloading stream optimized image efabb9a4-7907-4262-a138-86ad55c84b5b to [datastore1] OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e/OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e.vmdk on the data store datastore1 as vApp {{(pid=61857) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 963.021944] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Downloading image file data efabb9a4-7907-4262-a138-86ad55c84b5b to the ESX as VM named 'OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e' {{(pid=61857) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 963.105767] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 963.105767] env[61857]: value = "resgroup-9" [ 963.105767] env[61857]: _type = "ResourcePool" [ 963.105767] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 963.106099] env[61857]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-18191fd0-1d34-4a69-9593-b819015171b4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.129239] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lease: (returnval){ [ 963.129239] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5204aaef-f6f9-0321-f62e-63aee6d74634" [ 963.129239] env[61857]: _type = "HttpNfcLease" [ 963.129239] env[61857]: } obtained for vApp import into resource pool (val){ [ 963.129239] env[61857]: value = "resgroup-9" [ 963.129239] env[61857]: _type = "ResourcePool" [ 963.129239] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 963.129628] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the lease: (returnval){ [ 963.129628] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5204aaef-f6f9-0321-f62e-63aee6d74634" [ 963.129628] env[61857]: _type = "HttpNfcLease" [ 963.129628] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 963.136125] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.136125] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5204aaef-f6f9-0321-f62e-63aee6d74634" [ 963.136125] env[61857]: _type = "HttpNfcLease" [ 963.136125] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 963.177101] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951370, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066186} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.177437] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.178310] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f677273-89a8-4235-a971-c028e3048d69 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.204791] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] fd43f45b-0f54-46c6-94f9-ce28d3b15bb1/fd43f45b-0f54-46c6-94f9-ce28d3b15bb1.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.205150] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53d83c29-248e-4ecb-9a40-327f8c813578 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.226579] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for the task: (returnval){ [ 963.226579] env[61857]: value = "task-951372" [ 963.226579] env[61857]: _type = "Task" [ 963.226579] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.236231] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951372, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.282950] env[61857]: DEBUG nova.network.neutron [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Successfully updated port: 971c8b59-b659-4d32-9fba-17fdfffd9fe2 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.441023] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 963.441023] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-db48857d-cab9-425e-b7d7-475a1dbf205d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.449323] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 963.449323] env[61857]: value = "task-951373" [ 963.449323] env[61857]: _type = "Task" [ 963.449323] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.458023] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951373, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.492389] env[61857]: DEBUG nova.scheduler.client.report [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.637892] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.637892] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5204aaef-f6f9-0321-f62e-63aee6d74634" [ 963.637892] env[61857]: _type = "HttpNfcLease" [ 963.637892] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 963.638272] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 963.638272] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5204aaef-f6f9-0321-f62e-63aee6d74634" [ 963.638272] env[61857]: _type = "HttpNfcLease" [ 963.638272] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 963.638991] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9e4da3-40fb-46a8-85b0-8ae76cab2c0a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.646915] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526976fb-ac82-f3a2-2e55-27fece11e925/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 963.647118] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Creating HTTP connection to write to file with size = 31660032 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526976fb-ac82-f3a2-2e55-27fece11e925/disk-0.vmdk. {{(pid=61857) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 963.705929] env[61857]: DEBUG nova.compute.manager [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Received event network-vif-plugged-971c8b59-b659-4d32-9fba-17fdfffd9fe2 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.706161] env[61857]: DEBUG oslo_concurrency.lockutils [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] Acquiring lock "ce67a15a-8604-4523-a8fa-a34ccf9914da-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.706370] env[61857]: DEBUG oslo_concurrency.lockutils [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] Lock "ce67a15a-8604-4523-a8fa-a34ccf9914da-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.706542] env[61857]: DEBUG oslo_concurrency.lockutils [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] Lock "ce67a15a-8604-4523-a8fa-a34ccf9914da-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.706711] env[61857]: DEBUG nova.compute.manager [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] No waiting events found dispatching network-vif-plugged-971c8b59-b659-4d32-9fba-17fdfffd9fe2 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.706933] env[61857]: WARNING nova.compute.manager [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Received unexpected event network-vif-plugged-971c8b59-b659-4d32-9fba-17fdfffd9fe2 for instance with vm_state building and task_state spawning. [ 963.707162] env[61857]: DEBUG nova.compute.manager [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Received event network-changed-971c8b59-b659-4d32-9fba-17fdfffd9fe2 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.707332] env[61857]: DEBUG nova.compute.manager [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Refreshing instance network info cache due to event network-changed-971c8b59-b659-4d32-9fba-17fdfffd9fe2. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 963.707522] env[61857]: DEBUG oslo_concurrency.lockutils [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] Acquiring lock "refresh_cache-ce67a15a-8604-4523-a8fa-a34ccf9914da" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.707664] env[61857]: DEBUG oslo_concurrency.lockutils [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] Acquired lock "refresh_cache-ce67a15a-8604-4523-a8fa-a34ccf9914da" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.707823] env[61857]: DEBUG nova.network.neutron [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Refreshing network info cache for port 971c8b59-b659-4d32-9fba-17fdfffd9fe2 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 963.714632] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e88dd4a8-fedb-4b70-8661-7c2761841571 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.736601] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951372, 'name': ReconfigVM_Task, 'duration_secs': 0.282192} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.736897] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Reconfigured VM instance instance-0000005b to attach disk [datastore2] fd43f45b-0f54-46c6-94f9-ce28d3b15bb1/fd43f45b-0f54-46c6-94f9-ce28d3b15bb1.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.737573] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfd80866-235f-4880-9a3d-e3908d2295ff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.746842] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for the task: (returnval){ [ 963.746842] env[61857]: value = "task-951374" [ 963.746842] env[61857]: _type = "Task" [ 963.746842] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.753864] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951374, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.785877] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "refresh_cache-ce67a15a-8604-4523-a8fa-a34ccf9914da" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.960756] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951373, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.998247] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.799s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.003032] env[61857]: DEBUG oslo_concurrency.lockutils [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.224s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.021902] env[61857]: INFO nova.scheduler.client.report [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Deleted allocations for instance 825f5d87-76b1-4e4e-9b3d-4be419c7d323 [ 964.252400] env[61857]: DEBUG nova.network.neutron [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 964.259743] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951374, 'name': Rename_Task, 'duration_secs': 0.256166} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.261745] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 964.262075] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a9baa916-5052-43b2-8e28-585357b11ec7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.269215] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for the task: (returnval){ [ 964.269215] env[61857]: value = "task-951375" [ 964.269215] env[61857]: _type = "Task" [ 964.269215] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.280497] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951375, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.366679] env[61857]: DEBUG nova.network.neutron [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.462740] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951373, 'name': CloneVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.534181] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4a02039-7f2d-41c2-8c64-419868e9f931 tempest-InstanceActionsTestJSON-658461919 tempest-InstanceActionsTestJSON-658461919-project-member] Lock "825f5d87-76b1-4e4e-9b3d-4be419c7d323" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.396s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.686131] env[61857]: DEBUG nova.compute.manager [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Stashing vm_state: active {{(pid=61857) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 964.791162] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951375, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.826880] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3569b29-f5b2-4ba5-82a1-b89c50d1933b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.834888] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff2b2b0-6713-4288-8dec-4b61dec1df06 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.871839] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Completed reading data from the image iterator. {{(pid=61857) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 964.872102] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526976fb-ac82-f3a2-2e55-27fece11e925/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 964.873075] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79793f59-4e6e-4f0c-a689-edb6ef3ffd89 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.876010] env[61857]: DEBUG oslo_concurrency.lockutils [req-a73cbfae-8f4a-4849-84cf-17b17770183c req-3f810424-6b82-4ca2-ace0-90d092a30b8a service nova] Releasing lock "refresh_cache-ce67a15a-8604-4523-a8fa-a34ccf9914da" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.877361] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01c55b3-7b18-4058-868a-4a47898ca2b4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.879738] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "refresh_cache-ce67a15a-8604-4523-a8fa-a34ccf9914da" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.879898] env[61857]: DEBUG nova.network.neutron [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.888498] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750d7122-0dde-4e13-acf9-bb128cf29585 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.899020] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526976fb-ac82-f3a2-2e55-27fece11e925/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 964.899020] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526976fb-ac82-f3a2-2e55-27fece11e925/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 964.899020] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-8209a586-097b-41b0-af38-b94633c1ad6a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.908990] env[61857]: DEBUG nova.compute.provider_tree [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.961623] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951373, 'name': CloneVM_Task, 'duration_secs': 1.052555} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.961940] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Created linked-clone VM from snapshot [ 964.962747] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5345136-6bcf-4a55-80db-5e3aee2c2081 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.970079] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Uploading image f5089382-e2f7-4775-8fe7-4c9acaa926dc {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 964.995175] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 964.995175] env[61857]: value = "vm-214186" [ 964.995175] env[61857]: _type = "VirtualMachine" [ 964.995175] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 964.996499] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-16ac2df9-be47-445e-94ba-3812e7105f94 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.002324] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lease: (returnval){ [ 965.002324] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52831d52-fa64-846d-0269-7b41a88639cd" [ 965.002324] env[61857]: _type = "HttpNfcLease" [ 965.002324] env[61857]: } obtained for exporting VM: (result){ [ 965.002324] env[61857]: value = "vm-214186" [ 965.002324] env[61857]: _type = "VirtualMachine" [ 965.002324] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 965.002664] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the lease: (returnval){ [ 965.002664] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52831d52-fa64-846d-0269-7b41a88639cd" [ 965.002664] env[61857]: _type = "HttpNfcLease" [ 965.002664] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 965.008790] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 965.008790] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52831d52-fa64-846d-0269-7b41a88639cd" [ 965.008790] env[61857]: _type = "HttpNfcLease" [ 965.008790] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 965.082920] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526976fb-ac82-f3a2-2e55-27fece11e925/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 965.083188] env[61857]: INFO nova.virt.vmwareapi.images [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Downloaded image file data efabb9a4-7907-4262-a138-86ad55c84b5b [ 965.084105] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b54caff-5403-41ef-abf4-88150d76f72c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.104756] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a621b59-5c36-4c0f-9b6a-7da4a206fa3f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.141342] env[61857]: INFO nova.virt.vmwareapi.images [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] The imported VM was unregistered [ 965.143827] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Caching image {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 965.144834] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Creating directory with path [datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.144834] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16b8555d-121d-4239-bb41-35ae61a96881 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.155243] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Created directory with path [datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.155433] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e/OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e.vmdk to [datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b/efabb9a4-7907-4262-a138-86ad55c84b5b.vmdk. {{(pid=61857) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 965.155774] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e1e40ce2-2b78-402a-b406-b83678587f6f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.161650] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 965.161650] env[61857]: value = "task-951378" [ 965.161650] env[61857]: _type = "Task" [ 965.161650] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.169499] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951378, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.214546] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.279742] env[61857]: DEBUG oslo_vmware.api [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951375, 'name': PowerOnVM_Task, 'duration_secs': 0.523831} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.283095] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 965.283095] env[61857]: INFO nova.compute.manager [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Took 9.21 seconds to spawn the instance on the hypervisor. [ 965.283095] env[61857]: DEBUG nova.compute.manager [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.283095] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1106ea-598c-493f-b320-87f3dfd3b07e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.412039] env[61857]: DEBUG nova.scheduler.client.report [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.441771] env[61857]: DEBUG nova.network.neutron [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 965.514213] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 965.514213] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52831d52-fa64-846d-0269-7b41a88639cd" [ 965.514213] env[61857]: _type = "HttpNfcLease" [ 965.514213] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 965.514662] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 965.514662] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52831d52-fa64-846d-0269-7b41a88639cd" [ 965.514662] env[61857]: _type = "HttpNfcLease" [ 965.514662] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 965.515327] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548ba367-821a-4895-a09c-68401d5ed43a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.525104] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524dbda5-30b2-1db4-66ae-f8617a84b06b/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 965.525349] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524dbda5-30b2-1db4-66ae-f8617a84b06b/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 965.643514] env[61857]: DEBUG nova.network.neutron [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Updating instance_info_cache with network_info: [{"id": "971c8b59-b659-4d32-9fba-17fdfffd9fe2", "address": "fa:16:3e:fe:4e:62", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap971c8b59-b6", "ovs_interfaceid": "971c8b59-b659-4d32-9fba-17fdfffd9fe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.675194] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951378, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.678193] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cba67e50-b08e-431e-91ac-4ee4993cfc27 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.806236] env[61857]: INFO nova.compute.manager [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Took 23.28 seconds to build instance. [ 966.146688] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "refresh_cache-ce67a15a-8604-4523-a8fa-a34ccf9914da" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.147424] env[61857]: DEBUG nova.compute.manager [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Instance network_info: |[{"id": "971c8b59-b659-4d32-9fba-17fdfffd9fe2", "address": "fa:16:3e:fe:4e:62", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap971c8b59-b6", "ovs_interfaceid": "971c8b59-b659-4d32-9fba-17fdfffd9fe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 966.147747] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:4e:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '971c8b59-b659-4d32-9fba-17fdfffd9fe2', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.158147] env[61857]: DEBUG oslo.service.loopingcall [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.158488] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 966.158745] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3bc07f2-8e8b-46dc-9bf0-b2755fbdd48c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.187379] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951378, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.188910] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.188910] env[61857]: value = "task-951379" [ 966.188910] env[61857]: _type = "Task" [ 966.188910] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.201414] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951379, 'name': CreateVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.312844] env[61857]: DEBUG oslo_concurrency.lockutils [None req-085b65d5-5dbb-45bb-8397-7e0f4f4ba24d tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.802s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.427599] env[61857]: DEBUG oslo_concurrency.lockutils [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.424s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.432449] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.203s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.433347] env[61857]: DEBUG nova.objects.instance [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lazy-loading 'resources' on Instance uuid c821ec43-9462-4276-bdc9-d54e64c0a3cf {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.687882] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951378, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.701655] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951379, 'name': CreateVM_Task, 'duration_secs': 0.469086} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.702828] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 966.702993] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.703106] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.703792] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.705099] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9805524-7b3d-4c32-9e7b-7f510cb7e0b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.711030] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 966.711030] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52f3c2d5-edb8-3361-1400-8444393fdf9c" [ 966.711030] env[61857]: _type = "Task" [ 966.711030] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.722584] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f3c2d5-edb8-3361-1400-8444393fdf9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.880340] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquiring lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.880752] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.881062] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquiring lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.881384] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.881611] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.884123] env[61857]: INFO nova.compute.manager [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Terminating instance [ 966.886418] env[61857]: DEBUG nova.compute.manager [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 966.886684] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 966.887574] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e908e6c8-09bc-4c24-879f-dcb6e971d3f6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.897996] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 966.898431] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2548b13-3fe1-4caf-98df-5c54198e337b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.907389] env[61857]: DEBUG oslo_vmware.api [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for the task: (returnval){ [ 966.907389] env[61857]: value = "task-951380" [ 966.907389] env[61857]: _type = "Task" [ 966.907389] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.919539] env[61857]: DEBUG oslo_vmware.api [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951380, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.007290] env[61857]: INFO nova.scheduler.client.report [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleted allocation for migration c72d05c4-0844-486a-bf03-11222909c20c [ 967.188884] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951378, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.207421] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7f155e-fcaa-404d-8ea7-951f402018d7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.237437] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f3c2d5-edb8-3361-1400-8444393fdf9c, 'name': SearchDatastore_Task, 'duration_secs': 0.085751} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.243998] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.244635] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.245277] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.245668] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.246203] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.255155] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5575bee-40da-499b-ae3e-2a7f39be8bc5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.259573] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2bfbb6-91c9-4f50-9ed1-58d06471b32b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.059022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-581f925f-e739-465f-b744-0cfdbdad4ef2 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "18a632be-322f-48a9-becf-51ff3b735ad4" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 14.852s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.059022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "18a632be-322f-48a9-becf-51ff3b735ad4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.059022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "18a632be-322f-48a9-becf-51ff3b735ad4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.059022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "18a632be-322f-48a9-becf-51ff3b735ad4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.059022] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "18a632be-322f-48a9-becf-51ff3b735ad4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.059457] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "18a632be-322f-48a9-becf-51ff3b735ad4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.061191] env[61857]: INFO nova.compute.manager [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Terminating instance [ 968.066760] env[61857]: DEBUG nova.compute.manager [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.066976] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 968.076555] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b988874b-e8ee-460c-b689-1000221b9d3b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.079271] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d159b5e-afa8-4162-988c-dbf0b489a172 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.081927] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.082150] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 968.083430] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9399b8a7-d63e-4b25-be5b-3da7c6286ac7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.091851] env[61857]: DEBUG oslo_vmware.api [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951380, 'name': PowerOffVM_Task, 'duration_secs': 0.967875} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.100141] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 968.100389] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 968.100682] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 968.101413] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 968.101413] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523a8e9f-3c93-aebb-acf8-926d9e0cbf74" [ 968.101413] env[61857]: _type = "Task" [ 968.101413] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.101840] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951378, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.410814} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.102107] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f96bb872-0cec-47de-b86b-a2ee9f035e45 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.103704] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6fd4e3d-22de-44bd-8472-06ee6020b41c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.106132] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb026ee-6bbc-45fb-84e5-d62225a9838f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.110102] env[61857]: INFO nova.virt.vmwareapi.ds_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e/OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e.vmdk to [datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b/efabb9a4-7907-4262-a138-86ad55c84b5b.vmdk. [ 968.110352] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Cleaning up location [datastore1] OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 968.110536] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_3c2bafc3-65d0-4964-a38b-b2c2346c302e {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.113560] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fbefba28-4ac5-4e9e-976f-af0034e460c9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.126295] env[61857]: DEBUG nova.compute.provider_tree [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.133039] env[61857]: DEBUG oslo_vmware.api [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 968.133039] env[61857]: value = "task-951381" [ 968.133039] env[61857]: _type = "Task" [ 968.133039] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.133039] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 968.133039] env[61857]: value = "task-951383" [ 968.133039] env[61857]: _type = "Task" [ 968.133039] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.133402] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523a8e9f-3c93-aebb-acf8-926d9e0cbf74, 'name': SearchDatastore_Task, 'duration_secs': 0.009845} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.140313] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57efbdf5-5c57-4e5c-84b9-b7e8a4167a65 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.150399] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 968.150399] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524cccd7-6c54-2201-ef93-5271895fafc1" [ 968.150399] env[61857]: _type = "Task" [ 968.150399] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.150983] env[61857]: DEBUG oslo_vmware.api [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951381, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.154073] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951383, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.162702] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524cccd7-6c54-2201-ef93-5271895fafc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.200035] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 968.200337] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 968.200521] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Deleting the datastore file [datastore2] fd43f45b-0f54-46c6-94f9-ce28d3b15bb1 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.200840] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ba7414e-566b-44f5-9886-5e35e2ebef86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.209120] env[61857]: DEBUG oslo_vmware.api [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for the task: (returnval){ [ 968.209120] env[61857]: value = "task-951384" [ 968.209120] env[61857]: _type = "Task" [ 968.209120] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.217905] env[61857]: DEBUG oslo_vmware.api [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951384, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.311746] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.312076] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.635980] env[61857]: DEBUG nova.scheduler.client.report [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.651028] env[61857]: DEBUG oslo_vmware.api [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951381, 'name': PowerOffVM_Task, 'duration_secs': 0.19906} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.653775] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 968.653961] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 968.654234] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951383, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.058507} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.654503] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-942bc71d-b8bd-41b6-8e26-8f7c2e41b4ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.655950] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.656146] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "[datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b/efabb9a4-7907-4262-a138-86ad55c84b5b.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.656388] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b/efabb9a4-7907-4262-a138-86ad55c84b5b.vmdk to [datastore1] 781a2790-e317-46fe-9be8-40c9e5f1f771/781a2790-e317-46fe-9be8-40c9e5f1f771.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 968.659947] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f231637-0e66-4728-bb41-ca5cbcccf245 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.669124] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524cccd7-6c54-2201-ef93-5271895fafc1, 'name': SearchDatastore_Task, 'duration_secs': 0.019572} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.669232] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.669443] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] ce67a15a-8604-4523-a8fa-a34ccf9914da/ce67a15a-8604-4523-a8fa-a34ccf9914da.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 968.669748] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 968.669748] env[61857]: value = "task-951386" [ 968.669748] env[61857]: _type = "Task" [ 968.669748] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.669926] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e54dda5a-4b5e-4f93-8764-8cd202599d00 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.680796] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951386, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.682204] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 968.682204] env[61857]: value = "task-951387" [ 968.682204] env[61857]: _type = "Task" [ 968.682204] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.690687] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951387, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.718606] env[61857]: DEBUG oslo_vmware.api [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Task: {'id': task-951384, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193076} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.718858] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.719101] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 968.719309] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 968.719513] env[61857]: INFO nova.compute.manager [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Took 1.83 seconds to destroy the instance on the hypervisor. [ 968.719809] env[61857]: DEBUG oslo.service.loopingcall [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.720146] env[61857]: DEBUG nova.compute.manager [-] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 968.720326] env[61857]: DEBUG nova.network.neutron [-] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 968.747317] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 968.747596] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 968.747733] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleting the datastore file [datastore1] 18a632be-322f-48a9-becf-51ff3b735ad4 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.748013] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6151196-19c5-4531-b942-aaacaab90d02 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.754875] env[61857]: DEBUG oslo_vmware.api [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 968.754875] env[61857]: value = "task-951388" [ 968.754875] env[61857]: _type = "Task" [ 968.754875] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.763810] env[61857]: DEBUG oslo_vmware.api [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951388, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.814118] env[61857]: DEBUG nova.compute.manager [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 968.992482] env[61857]: DEBUG nova.compute.manager [req-5743d0af-66c9-4064-b378-b6ec33575ab2 req-ec705c70-c6b5-44ec-aef9-b1c90c66a612 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Received event network-vif-deleted-9c8f3dfe-a913-4426-9e25-9a2c4786fd82 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 968.992733] env[61857]: INFO nova.compute.manager [req-5743d0af-66c9-4064-b378-b6ec33575ab2 req-ec705c70-c6b5-44ec-aef9-b1c90c66a612 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Neutron deleted interface 9c8f3dfe-a913-4426-9e25-9a2c4786fd82; detaching it from the instance and deleting it from the info cache [ 968.992916] env[61857]: DEBUG nova.network.neutron [req-5743d0af-66c9-4064-b378-b6ec33575ab2 req-ec705c70-c6b5-44ec-aef9-b1c90c66a612 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.146617] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.714s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.150988] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.511s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.150988] env[61857]: DEBUG nova.objects.instance [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lazy-loading 'resources' on Instance uuid 0de8e67e-24b9-48d0-ac90-b5ec5b93e243 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.181271] env[61857]: INFO nova.scheduler.client.report [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Deleted allocations for instance c821ec43-9462-4276-bdc9-d54e64c0a3cf [ 969.197311] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951386, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.206084] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951387, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.269789] env[61857]: DEBUG oslo_vmware.api [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951388, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.337468] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.466633] env[61857]: DEBUG nova.network.neutron [-] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.496876] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-84e99f18-39ab-462c-9f0c-7d5759ab3a1a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.508539] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd21202-9b44-4303-8f31-2fe65cd2ccc4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.541552] env[61857]: DEBUG nova.compute.manager [req-5743d0af-66c9-4064-b378-b6ec33575ab2 req-ec705c70-c6b5-44ec-aef9-b1c90c66a612 service nova] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Detach interface failed, port_id=9c8f3dfe-a913-4426-9e25-9a2c4786fd82, reason: Instance fd43f45b-0f54-46c6-94f9-ce28d3b15bb1 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 969.690141] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951386, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.700924] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951387, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.92806} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.701975] env[61857]: DEBUG oslo_concurrency.lockutils [None req-eb95b197-f037-4853-a273-3710c1ae0f55 tempest-ServerAddressesNegativeTestJSON-1593364250 tempest-ServerAddressesNegativeTestJSON-1593364250-project-member] Lock "c821ec43-9462-4276-bdc9-d54e64c0a3cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.154s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.702633] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] ce67a15a-8604-4523-a8fa-a34ccf9914da/ce67a15a-8604-4523-a8fa-a34ccf9914da.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 969.702869] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 969.703625] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0299c13d-9d1f-4662-b11d-400f9e4ff674 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.714058] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 969.714058] env[61857]: value = "task-951389" [ 969.714058] env[61857]: _type = "Task" [ 969.714058] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.728365] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951389, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.769284] env[61857]: DEBUG oslo_vmware.api [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951388, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.985246} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.772753] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.773055] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 969.773235] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 969.773428] env[61857]: INFO nova.compute.manager [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Took 1.71 seconds to destroy the instance on the hypervisor. [ 969.773702] env[61857]: DEBUG oslo.service.loopingcall [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.774162] env[61857]: DEBUG nova.compute.manager [-] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.774276] env[61857]: DEBUG nova.network.neutron [-] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 969.896509] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3dbdaa-98f5-41dc-af25-ea0f88ffc5c5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.905959] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff82340-8ef3-4ba4-967a-9962e097ebb2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.939044] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4862bce-eca4-4d12-8770-57ed0667c780 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.947612] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f7aff0d-6e07-4f8b-8478-4c2c5dad7387 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.962981] env[61857]: DEBUG nova.compute.provider_tree [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.970489] env[61857]: INFO nova.compute.manager [-] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Took 1.25 seconds to deallocate network for instance. [ 970.183880] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951386, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.224031] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951389, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094774} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.224031] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.227572] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835d3829-cb0c-4e9f-bce2-b5c7c096b3cc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.250375] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] ce67a15a-8604-4523-a8fa-a34ccf9914da/ce67a15a-8604-4523-a8fa-a34ccf9914da.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.250876] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f6b60645-62cd-40ca-af73-50095ccc6218 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.251138] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f6b60645-62cd-40ca-af73-50095ccc6218 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.251537] env[61857]: DEBUG nova.objects.instance [None req-f6b60645-62cd-40ca-af73-50095ccc6218 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'flavor' on Instance uuid f982f0c5-c701-4e5b-b96d-c21a3c0da24f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.253019] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2bf58b8-8876-4dd8-8461-80573083fc09 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.276202] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 970.276202] env[61857]: value = "task-951390" [ 970.276202] env[61857]: _type = "Task" [ 970.276202] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.285285] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951390, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.466429] env[61857]: DEBUG nova.scheduler.client.report [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.479237] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.521993] env[61857]: DEBUG nova.network.neutron [-] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.683662] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951386, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.770776] env[61857]: DEBUG nova.objects.instance [None req-f6b60645-62cd-40ca-af73-50095ccc6218 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'pci_requests' on Instance uuid f982f0c5-c701-4e5b-b96d-c21a3c0da24f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.791783] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951390, 'name': ReconfigVM_Task, 'duration_secs': 0.39225} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.792145] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Reconfigured VM instance instance-0000005c to attach disk [datastore1] ce67a15a-8604-4523-a8fa-a34ccf9914da/ce67a15a-8604-4523-a8fa-a34ccf9914da.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.792827] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68e310c1-aa26-4a57-9aff-9c9d9a7a21bf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.803500] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 970.803500] env[61857]: value = "task-951391" [ 970.803500] env[61857]: _type = "Task" [ 970.803500] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.812243] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951391, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.972228] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.822s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.975523] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.614s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.975523] env[61857]: DEBUG nova.objects.instance [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lazy-loading 'resources' on Instance uuid ce5f713d-8b63-4b95-a5ad-feddb95f921d {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.006772] env[61857]: INFO nova.scheduler.client.report [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Deleted allocations for instance 0de8e67e-24b9-48d0-ac90-b5ec5b93e243 [ 971.029598] env[61857]: INFO nova.compute.manager [-] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Took 1.25 seconds to deallocate network for instance. [ 971.188062] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951386, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.276552] env[61857]: DEBUG nova.objects.base [None req-f6b60645-62cd-40ca-af73-50095ccc6218 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 971.276825] env[61857]: DEBUG nova.network.neutron [None req-f6b60645-62cd-40ca-af73-50095ccc6218 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 971.314838] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951391, 'name': Rename_Task, 'duration_secs': 0.172905} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.315282] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 971.315522] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-697f1e24-300f-46fd-b042-737dc749466c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.322311] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 971.322311] env[61857]: value = "task-951392" [ 971.322311] env[61857]: _type = "Task" [ 971.322311] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.331424] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.380044] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f6b60645-62cd-40ca-af73-50095ccc6218 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.129s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.419674] env[61857]: DEBUG nova.compute.manager [req-023897a2-cf03-453a-8020-00ccc869924a req-a96e3a5f-1ad1-4d26-b0fc-1adb129d1546 service nova] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Received event network-vif-deleted-e4c39a04-24d7-48d1-a089-2304778bd059 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 971.519208] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9711dd2c-3d50-48c5-b967-37b8cd72e264 tempest-AttachVolumeNegativeTest-141958123 tempest-AttachVolumeNegativeTest-141958123-project-member] Lock "0de8e67e-24b9-48d0-ac90-b5ec5b93e243" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.903s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.540915] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.691522] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951386, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.616516} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.691883] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/efabb9a4-7907-4262-a138-86ad55c84b5b/efabb9a4-7907-4262-a138-86ad55c84b5b.vmdk to [datastore1] 781a2790-e317-46fe-9be8-40c9e5f1f771/781a2790-e317-46fe-9be8-40c9e5f1f771.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 971.693013] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328fe2e5-6b49-4c50-addf-5f2a0a1e0401 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.719875] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 781a2790-e317-46fe-9be8-40c9e5f1f771/781a2790-e317-46fe-9be8-40c9e5f1f771.vmdk or device None with type streamOptimized {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.723636] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a695aad-1fb0-481a-b26c-bf8c993ef511 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.746034] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 971.746034] env[61857]: value = "task-951393" [ 971.746034] env[61857]: _type = "Task" [ 971.746034] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.759795] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951393, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.781342] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fee191-5527-43d3-94db-40bddd997233 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.789390] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687641d6-6d67-4c52-9cd4-97cf948cec4d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.829324] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2aede9b-b3f3-4edc-900e-9d4768c019cd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.837266] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951392, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.840470] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc37606d-d00a-4ecb-8b3c-41bf1c102dfa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.853748] env[61857]: DEBUG nova.compute.provider_tree [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.259279] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951393, 'name': ReconfigVM_Task, 'duration_secs': 0.306497} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.259626] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 781a2790-e317-46fe-9be8-40c9e5f1f771/781a2790-e317-46fe-9be8-40c9e5f1f771.vmdk or device None with type streamOptimized {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.260258] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a367fda3-93c7-486d-91bb-f7c76c4eb9a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.267720] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 972.267720] env[61857]: value = "task-951395" [ 972.267720] env[61857]: _type = "Task" [ 972.267720] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.276362] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951395, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.334890] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951392, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.357833] env[61857]: DEBUG nova.scheduler.client.report [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.779354] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951395, 'name': Rename_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.835526] env[61857]: DEBUG oslo_vmware.api [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951392, 'name': PowerOnVM_Task, 'duration_secs': 1.147003} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.835800] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 972.836018] env[61857]: INFO nova.compute.manager [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Took 10.53 seconds to spawn the instance on the hypervisor. [ 972.836204] env[61857]: DEBUG nova.compute.manager [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.836956] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb5196a-ae8f-4559-8c35-54aaf46f52cc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.864515] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.889s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.866792] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 7.652s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.889159] env[61857]: INFO nova.scheduler.client.report [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Deleted allocations for instance ce5f713d-8b63-4b95-a5ad-feddb95f921d [ 973.279937] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951395, 'name': Rename_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.284549] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.284824] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.285224] env[61857]: DEBUG nova.objects.instance [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'flavor' on Instance uuid f982f0c5-c701-4e5b-b96d-c21a3c0da24f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.353943] env[61857]: INFO nova.compute.manager [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Took 22.80 seconds to build instance. [ 973.371087] env[61857]: INFO nova.compute.claims [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 973.396234] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45cd4068-d061-417c-ba4c-af9459808ad2 tempest-ImagesTestJSON-1046725449 tempest-ImagesTestJSON-1046725449-project-member] Lock "ce5f713d-8b63-4b95-a5ad-feddb95f921d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.893s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.780137] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951395, 'name': Rename_Task, 'duration_secs': 1.156282} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.780464] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 973.780736] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6cd9939-8aa1-4c87-9f93-285af1c2240f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.790384] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 973.790384] env[61857]: value = "task-951396" [ 973.790384] env[61857]: _type = "Task" [ 973.790384] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.803142] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951396, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.856737] env[61857]: DEBUG oslo_concurrency.lockutils [None req-17319ddf-ebee-4f81-abb0-05d7436354ee tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "ce67a15a-8604-4523-a8fa-a34ccf9914da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.307s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.877362] env[61857]: INFO nova.compute.resource_tracker [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating resource usage from migration 29cccd98-cdb2-46c7-920c-198017f1584b [ 973.895994] env[61857]: DEBUG nova.objects.instance [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'pci_requests' on Instance uuid f982f0c5-c701-4e5b-b96d-c21a3c0da24f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.109116] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5b330b-7dc4-4518-b548-14ff9befb55b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.116978] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e58e54b-df88-401b-8dd6-09f29a587f3c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.148598] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0b703a-118e-4152-b7fb-164d1cc34cf8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.156462] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e67d9d7-b1b8-4bba-9728-496fb8281b05 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.170381] env[61857]: DEBUG nova.compute.provider_tree [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.300825] env[61857]: DEBUG oslo_vmware.api [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951396, 'name': PowerOnVM_Task, 'duration_secs': 0.463811} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.301214] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 974.398898] env[61857]: DEBUG nova.objects.base [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 974.399113] env[61857]: DEBUG nova.network.neutron [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 974.414455] env[61857]: DEBUG nova.compute.manager [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.415427] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18073089-fffc-4187-bb08-13f646f78b5b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.461119] env[61857]: DEBUG nova.policy [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9056d344ff614950a60983450a2a40ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5653ed6ab0a948ef89bb03cf7426dd10', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 974.673659] env[61857]: DEBUG nova.scheduler.client.report [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.735298] env[61857]: DEBUG nova.network.neutron [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Successfully created port: d306c021-073e-45f7-9916-f3a6bc7a7dff {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.936824] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4d18ff6-0684-418a-8c8a-023f2478f06e tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 26.105s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.134878] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.135188] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.180815] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.314s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.180929] env[61857]: INFO nova.compute.manager [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Migrating [ 975.188145] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.851s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.190191] env[61857]: INFO nova.compute.claims [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 975.486268] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524dbda5-30b2-1db4-66ae-f8617a84b06b/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 975.487442] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e509ce7d-24ca-48b0-a731-a91d531640d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.495575] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524dbda5-30b2-1db4-66ae-f8617a84b06b/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 975.495768] env[61857]: ERROR oslo_vmware.rw_handles [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524dbda5-30b2-1db4-66ae-f8617a84b06b/disk-0.vmdk due to incomplete transfer. [ 975.496049] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-bceaab67-aafb-49aa-93b6-e16260d5a7b7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.503130] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524dbda5-30b2-1db4-66ae-f8617a84b06b/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 975.503343] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Uploaded image f5089382-e2f7-4775-8fe7-4c9acaa926dc to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 975.507708] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 975.507708] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-76c64701-2aca-481d-81ab-449d34fd331f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.513714] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 975.513714] env[61857]: value = "task-951397" [ 975.513714] env[61857]: _type = "Task" [ 975.513714] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.523712] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951397, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.641986] env[61857]: DEBUG nova.compute.manager [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 975.702687] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.702870] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.703585] env[61857]: DEBUG nova.network.neutron [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 976.023930] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951397, 'name': Destroy_Task} progress is 33%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.175509] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.355833] env[61857]: DEBUG nova.network.neutron [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Successfully updated port: d306c021-073e-45f7-9916-f3a6bc7a7dff {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 976.447021] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa37f816-ca30-4b59-8ff6-a79eb19e607a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.457097] env[61857]: DEBUG nova.compute.manager [req-fe215525-34a9-4ecd-abf5-d1a1d84b8daf req-2b0d5d18-b8dc-45c2-b9d7-809e1c59b04b service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received event network-vif-plugged-d306c021-073e-45f7-9916-f3a6bc7a7dff {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.457097] env[61857]: DEBUG oslo_concurrency.lockutils [req-fe215525-34a9-4ecd-abf5-d1a1d84b8daf req-2b0d5d18-b8dc-45c2-b9d7-809e1c59b04b service nova] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.457097] env[61857]: DEBUG oslo_concurrency.lockutils [req-fe215525-34a9-4ecd-abf5-d1a1d84b8daf req-2b0d5d18-b8dc-45c2-b9d7-809e1c59b04b service nova] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.457860] env[61857]: DEBUG oslo_concurrency.lockutils [req-fe215525-34a9-4ecd-abf5-d1a1d84b8daf req-2b0d5d18-b8dc-45c2-b9d7-809e1c59b04b service nova] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.457860] env[61857]: DEBUG nova.compute.manager [req-fe215525-34a9-4ecd-abf5-d1a1d84b8daf req-2b0d5d18-b8dc-45c2-b9d7-809e1c59b04b service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] No waiting events found dispatching network-vif-plugged-d306c021-073e-45f7-9916-f3a6bc7a7dff {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.457860] env[61857]: WARNING nova.compute.manager [req-fe215525-34a9-4ecd-abf5-d1a1d84b8daf req-2b0d5d18-b8dc-45c2-b9d7-809e1c59b04b service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received unexpected event network-vif-plugged-d306c021-073e-45f7-9916-f3a6bc7a7dff for instance with vm_state active and task_state None. [ 976.462340] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246e7fa4-fb00-4a42-82b8-89ebc5027023 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.493081] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1f6518-4399-41c0-aa68-70edd93cd04f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.500689] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfac493-61f4-42c9-b106-a02a27b86094 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.514196] env[61857]: DEBUG nova.compute.provider_tree [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.523368] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951397, 'name': Destroy_Task, 'duration_secs': 0.860638} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.523474] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Destroyed the VM [ 976.523701] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 976.524469] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b1fe13f1-15c1-46a2-9e0b-3584eef28317 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.530643] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 976.530643] env[61857]: value = "task-951399" [ 976.530643] env[61857]: _type = "Task" [ 976.530643] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.534533] env[61857]: DEBUG nova.network.neutron [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance_info_cache with network_info: [{"id": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "address": "fa:16:3e:32:96:34", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c3f890-dd", "ovs_interfaceid": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.542719] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951399, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.858806] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.859034] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.859260] env[61857]: DEBUG nova.network.neutron [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.020027] env[61857]: DEBUG nova.scheduler.client.report [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.037227] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.044136] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951399, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.248963] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6be2a40-abef-4066-b114-abf7b7fd7dfe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.256145] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-33ba970a-83f5-4f53-9bce-1b346112b151 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Suspending the VM {{(pid=61857) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 977.256418] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-3c168e28-2809-4828-9878-7afaa21a1ed0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.262139] env[61857]: DEBUG oslo_vmware.api [None req-33ba970a-83f5-4f53-9bce-1b346112b151 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 977.262139] env[61857]: value = "task-951400" [ 977.262139] env[61857]: _type = "Task" [ 977.262139] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.269965] env[61857]: DEBUG oslo_vmware.api [None req-33ba970a-83f5-4f53-9bce-1b346112b151 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951400, 'name': SuspendVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.403593] env[61857]: WARNING nova.network.neutron [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] 30b7d27c-744c-46a6-8769-b1b1c95306e4 already exists in list: networks containing: ['30b7d27c-744c-46a6-8769-b1b1c95306e4']. ignoring it [ 977.527671] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.528329] env[61857]: DEBUG nova.compute.manager [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 977.531608] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.053s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.532251] env[61857]: DEBUG nova.objects.instance [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lazy-loading 'resources' on Instance uuid fd43f45b-0f54-46c6-94f9-ce28d3b15bb1 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.551111] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951399, 'name': RemoveSnapshot_Task, 'duration_secs': 0.603603} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.551434] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 977.551763] env[61857]: DEBUG nova.compute.manager [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.553170] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8347544-ed3b-4c81-9a67-bc1e7e1e21ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.719824] env[61857]: DEBUG nova.network.neutron [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d306c021-073e-45f7-9916-f3a6bc7a7dff", "address": "fa:16:3e:52:3f:1e", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd306c021-07", "ovs_interfaceid": "d306c021-073e-45f7-9916-f3a6bc7a7dff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.772790] env[61857]: DEBUG oslo_vmware.api [None req-33ba970a-83f5-4f53-9bce-1b346112b151 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951400, 'name': SuspendVM_Task} progress is 70%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.038383] env[61857]: DEBUG nova.compute.utils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.040040] env[61857]: DEBUG nova.compute.manager [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 978.040320] env[61857]: DEBUG nova.network.neutron [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 978.076319] env[61857]: INFO nova.compute.manager [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Shelve offloading [ 978.081019] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 978.081019] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9c300d0-ae8d-4d86-b2f1-a7b17143eca3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.086648] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 978.086648] env[61857]: value = "task-951401" [ 978.086648] env[61857]: _type = "Task" [ 978.086648] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.090474] env[61857]: DEBUG nova.policy [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac8387df3290404283263440672de653', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3bbcf7c2937040e1906e2273a07b671b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.097879] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951401, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.225458] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.225948] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.226130] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.227079] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef531dc2-0fe3-49c1-a593-a4ed909698c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.246165] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 978.247394] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 978.247394] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 978.247394] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 978.247394] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 978.247849] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 978.247849] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 978.248011] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 978.248111] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 978.248277] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 978.248866] env[61857]: DEBUG nova.virt.hardware [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 978.255434] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Reconfiguring VM to attach interface {{(pid=61857) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 978.258640] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9789d589-9d7c-4a95-9fa7-6ce043514565 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.279444] env[61857]: DEBUG oslo_vmware.api [None req-33ba970a-83f5-4f53-9bce-1b346112b151 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951400, 'name': SuspendVM_Task, 'duration_secs': 0.586938} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.281239] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-33ba970a-83f5-4f53-9bce-1b346112b151 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Suspended the VM {{(pid=61857) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 978.281525] env[61857]: DEBUG nova.compute.manager [None req-33ba970a-83f5-4f53-9bce-1b346112b151 tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.281939] env[61857]: DEBUG oslo_vmware.api [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 978.281939] env[61857]: value = "task-951402" [ 978.281939] env[61857]: _type = "Task" [ 978.281939] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.282927] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48321fa6-4614-49ce-9d8b-fa81c0197635 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.296832] env[61857]: DEBUG oslo_vmware.api [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951402, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.334424] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8889dc03-81d9-45de-ba79-88b4cfcc1097 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.343596] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cae6721-16a2-42a4-a770-4befaad8afb4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.376932] env[61857]: DEBUG nova.network.neutron [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Successfully created port: 7a4acc68-8312-44db-90f7-f628627266cc {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.379447] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c043086-70aa-4ce9-a387-8626c93a21a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.387738] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699d2bd4-9c85-4a71-97ef-b3f259fc2c6c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.405037] env[61857]: DEBUG nova.compute.provider_tree [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.546649] env[61857]: DEBUG nova.compute.manager [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 978.564524] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a58e490-313c-4774-be6b-b2965c0860b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.587665] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance '71920a62-090d-4df1-937e-87df3b043e28' progress to 0 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 978.602382] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] VM already powered off {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 978.602636] env[61857]: DEBUG nova.compute.manager [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.603487] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a55328f-fbdb-4a9a-936b-dab520c4b360 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.611841] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.612116] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.612217] env[61857]: DEBUG nova.network.neutron [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 978.686969] env[61857]: DEBUG oslo_concurrency.lockutils [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.687270] env[61857]: DEBUG oslo_concurrency.lockutils [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.787204] env[61857]: DEBUG nova.compute.manager [req-2d6b1c76-4e40-4021-8056-3349ac6fbdc9 req-6ee0b57d-49fa-4f40-b72f-18bf058396ba service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received event network-changed-d306c021-073e-45f7-9916-f3a6bc7a7dff {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 978.787467] env[61857]: DEBUG nova.compute.manager [req-2d6b1c76-4e40-4021-8056-3349ac6fbdc9 req-6ee0b57d-49fa-4f40-b72f-18bf058396ba service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Refreshing instance network info cache due to event network-changed-d306c021-073e-45f7-9916-f3a6bc7a7dff. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 978.787733] env[61857]: DEBUG oslo_concurrency.lockutils [req-2d6b1c76-4e40-4021-8056-3349ac6fbdc9 req-6ee0b57d-49fa-4f40-b72f-18bf058396ba service nova] Acquiring lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.787922] env[61857]: DEBUG oslo_concurrency.lockutils [req-2d6b1c76-4e40-4021-8056-3349ac6fbdc9 req-6ee0b57d-49fa-4f40-b72f-18bf058396ba service nova] Acquired lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.788977] env[61857]: DEBUG nova.network.neutron [req-2d6b1c76-4e40-4021-8056-3349ac6fbdc9 req-6ee0b57d-49fa-4f40-b72f-18bf058396ba service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Refreshing network info cache for port d306c021-073e-45f7-9916-f3a6bc7a7dff {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 978.803115] env[61857]: DEBUG oslo_vmware.api [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951402, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.906976] env[61857]: DEBUG nova.scheduler.client.report [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.096124] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 979.096598] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9971ddfc-6978-4f5f-87ed-4e780a6926ca {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.107245] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 979.107245] env[61857]: value = "task-951403" [ 979.107245] env[61857]: _type = "Task" [ 979.107245] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.118979] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951403, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.190763] env[61857]: INFO nova.compute.manager [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Detaching volume 39d32790-20cc-4af1-a460-e9e10cb159b3 [ 979.235198] env[61857]: INFO nova.virt.block_device [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Attempting to driver detach volume 39d32790-20cc-4af1-a460-e9e10cb159b3 from mountpoint /dev/sdb [ 979.235479] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Volume detach. Driver type: vmdk {{(pid=61857) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 979.235667] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214162', 'volume_id': '39d32790-20cc-4af1-a460-e9e10cb159b3', 'name': 'volume-39d32790-20cc-4af1-a460-e9e10cb159b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd9d05629-79ef-45c6-ac54-b3bc18d306f4', 'attached_at': '', 'detached_at': '', 'volume_id': '39d32790-20cc-4af1-a460-e9e10cb159b3', 'serial': '39d32790-20cc-4af1-a460-e9e10cb159b3'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 979.236906] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d0b0ab-f7ab-4554-83f9-aa64b2d20e37 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.263393] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db1c50d-0ad3-443c-a1de-7a640b4199f3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.270955] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55e9d41-5f67-4ad5-bee6-399754158eb2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.301851] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f55c735-b2b6-4c54-8584-ff34bc8e1f89 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.310827] env[61857]: DEBUG oslo_vmware.api [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951402, 'name': ReconfigVM_Task, 'duration_secs': 0.685953} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.321935] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.321935] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Reconfigured VM to attach interface {{(pid=61857) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 979.328835] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] The volume has not been displaced from its original location: [datastore2] volume-39d32790-20cc-4af1-a460-e9e10cb159b3/volume-39d32790-20cc-4af1-a460-e9e10cb159b3.vmdk. No consolidation needed. {{(pid=61857) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 979.329944] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 979.330500] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34bfad9e-41fc-4c79-a4fb-30fc6094dc22 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.355154] env[61857]: DEBUG oslo_vmware.api [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 979.355154] env[61857]: value = "task-951404" [ 979.355154] env[61857]: _type = "Task" [ 979.355154] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.361790] env[61857]: DEBUG oslo_vmware.api [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951404, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.411338] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.879s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.414144] env[61857]: DEBUG nova.network.neutron [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Updating instance_info_cache with network_info: [{"id": "5bd8b89d-fe07-48dc-a773-c222f7296a19", "address": "fa:16:3e:71:32:7d", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bd8b89d-fe", "ovs_interfaceid": "5bd8b89d-fe07-48dc-a773-c222f7296a19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.415479] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.875s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.415890] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.417358] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.242s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.419560] env[61857]: INFO nova.compute.claims [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.450226] env[61857]: INFO nova.scheduler.client.report [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Deleted allocations for instance fd43f45b-0f54-46c6-94f9-ce28d3b15bb1 [ 979.475318] env[61857]: INFO nova.scheduler.client.report [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleted allocations for instance 18a632be-322f-48a9-becf-51ff3b735ad4 [ 979.558265] env[61857]: DEBUG nova.compute.manager [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 979.584253] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.584524] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.584689] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.584875] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.585039] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.585199] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.585410] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.585577] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.585790] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.585909] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.586125] env[61857]: DEBUG nova.virt.hardware [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.587007] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c9ea5a-7a86-4a0a-a28b-22aa86455e7d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.590229] env[61857]: DEBUG nova.network.neutron [req-2d6b1c76-4e40-4021-8056-3349ac6fbdc9 req-6ee0b57d-49fa-4f40-b72f-18bf058396ba service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updated VIF entry in instance network info cache for port d306c021-073e-45f7-9916-f3a6bc7a7dff. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 979.590658] env[61857]: DEBUG nova.network.neutron [req-2d6b1c76-4e40-4021-8056-3349ac6fbdc9 req-6ee0b57d-49fa-4f40-b72f-18bf058396ba service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d306c021-073e-45f7-9916-f3a6bc7a7dff", "address": "fa:16:3e:52:3f:1e", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd306c021-07", "ovs_interfaceid": "d306c021-073e-45f7-9916-f3a6bc7a7dff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.600822] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24b0eda-61cd-414b-8717-61bcb4170dfa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.624770] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951403, 'name': PowerOffVM_Task, 'duration_secs': 0.19758} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.625129] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 979.625292] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance '71920a62-090d-4df1-937e-87df3b043e28' progress to 17 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 979.833301] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d9c7106d-5e68-4f1e-9c8a-0257fed4fc99 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.548s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.864022] env[61857]: DEBUG oslo_vmware.api [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951404, 'name': ReconfigVM_Task, 'duration_secs': 0.221807} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.864314] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 979.870444] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04ebfe89-ce5e-4b6d-83ee-e362732f0131 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.886610] env[61857]: DEBUG oslo_vmware.api [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 979.886610] env[61857]: value = "task-951405" [ 979.886610] env[61857]: _type = "Task" [ 979.886610] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.895447] env[61857]: DEBUG oslo_vmware.api [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951405, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.927296] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.958093] env[61857]: DEBUG nova.network.neutron [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Successfully updated port: 7a4acc68-8312-44db-90f7-f628627266cc {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.959437] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ff776dd5-cdd8-4d30-b516-d6e0a5863c76 tempest-ServerPasswordTestJSON-1515969797 tempest-ServerPasswordTestJSON-1515969797-project-member] Lock "fd43f45b-0f54-46c6-94f9-ce28d3b15bb1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.079s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.987124] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9fe8eeb-d255-4548-8e41-f7e3610dbad7 tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "18a632be-322f-48a9-becf-51ff3b735ad4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.928s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.056570] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquiring lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.056893] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.098661] env[61857]: DEBUG oslo_concurrency.lockutils [req-2d6b1c76-4e40-4021-8056-3349ac6fbdc9 req-6ee0b57d-49fa-4f40-b72f-18bf058396ba service nova] Releasing lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.131452] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.131954] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.131954] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.132151] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.132306] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.132462] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.132668] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.132829] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.132997] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.133527] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.133779] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.141549] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b53a4b01-fabf-480a-af29-c6172b714b90 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.166569] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 980.166569] env[61857]: value = "task-951406" [ 980.166569] env[61857]: _type = "Task" [ 980.166569] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.177554] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951406, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.398684] env[61857]: DEBUG oslo_vmware.api [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951405, 'name': ReconfigVM_Task, 'duration_secs': 0.155364} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.399011] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214162', 'volume_id': '39d32790-20cc-4af1-a460-e9e10cb159b3', 'name': 'volume-39d32790-20cc-4af1-a460-e9e10cb159b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd9d05629-79ef-45c6-ac54-b3bc18d306f4', 'attached_at': '', 'detached_at': '', 'volume_id': '39d32790-20cc-4af1-a460-e9e10cb159b3', 'serial': '39d32790-20cc-4af1-a460-e9e10cb159b3'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 980.462196] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "refresh_cache-2cd4f38e-4111-45b2-a8bd-dc0d38ada293" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.462196] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "refresh_cache-2cd4f38e-4111-45b2-a8bd-dc0d38ada293" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.462196] env[61857]: DEBUG nova.network.neutron [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 980.560064] env[61857]: DEBUG nova.compute.manager [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 980.641017] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1505b8c8-d9da-41ab-a3e4-db0ebe6c23c2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.647044] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6469a0dd-eddf-4e4c-bc8d-f11f9b548cac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.685112] env[61857]: INFO nova.compute.manager [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Resuming [ 980.685112] env[61857]: DEBUG nova.objects.instance [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lazy-loading 'flavor' on Instance uuid 781a2790-e317-46fe-9be8-40c9e5f1f771 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.690363] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694fa3aa-d022-4f5e-a902-4ae0cfc133ab {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.701045] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951406, 'name': ReconfigVM_Task, 'duration_secs': 0.177018} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.703246] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance '71920a62-090d-4df1-937e-87df3b043e28' progress to 33 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 980.708488] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc071f17-f3a8-451d-9c64-5423ae62ac3f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.722857] env[61857]: DEBUG nova.compute.provider_tree [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.733633] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 980.734684] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc63d425-8947-4514-a4bf-058f5b504ebc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.741927] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 980.742136] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b161b4ca-f2db-4c6c-b036-088484fc4d8b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.819075] env[61857]: DEBUG nova.compute.manager [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Received event network-vif-plugged-7a4acc68-8312-44db-90f7-f628627266cc {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 980.819299] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Acquiring lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.819491] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.819668] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.819839] env[61857]: DEBUG nova.compute.manager [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] No waiting events found dispatching network-vif-plugged-7a4acc68-8312-44db-90f7-f628627266cc {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 980.820013] env[61857]: WARNING nova.compute.manager [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Received unexpected event network-vif-plugged-7a4acc68-8312-44db-90f7-f628627266cc for instance with vm_state building and task_state spawning. [ 980.820333] env[61857]: DEBUG nova.compute.manager [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Received event network-changed-7a4acc68-8312-44db-90f7-f628627266cc {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 980.820516] env[61857]: DEBUG nova.compute.manager [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Refreshing instance network info cache due to event network-changed-7a4acc68-8312-44db-90f7-f628627266cc. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 980.820767] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Acquiring lock "refresh_cache-2cd4f38e-4111-45b2-a8bd-dc0d38ada293" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.848597] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 980.848597] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 980.848597] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleting the datastore file [datastore2] c3c9df36-d523-4a0f-9484-4446ad8c4138 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.848597] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4fa3170-7159-4344-9e2e-42bc741c6649 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.856194] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 980.856194] env[61857]: value = "task-951408" [ 980.856194] env[61857]: _type = "Task" [ 980.856194] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.863429] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951408, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.960408] env[61857]: DEBUG nova.objects.instance [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lazy-loading 'flavor' on Instance uuid d9d05629-79ef-45c6-ac54-b3bc18d306f4 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.049248] env[61857]: DEBUG nova.network.neutron [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 981.087072] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.215047] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 981.215309] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 981.215476] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.215666] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 981.215820] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.215972] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 981.216585] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 981.216829] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 981.217142] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 981.217332] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 981.217691] env[61857]: DEBUG nova.virt.hardware [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.223732] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Reconfiguring VM instance instance-00000035 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 981.227058] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-662b4eb7-7b20-4df7-8840-0663fb62266c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.240359] env[61857]: DEBUG nova.scheduler.client.report [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.250346] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 981.250346] env[61857]: value = "task-951409" [ 981.250346] env[61857]: _type = "Task" [ 981.250346] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.258555] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951409, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.368672] env[61857]: DEBUG oslo_vmware.api [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951408, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144262} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.369496] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.369496] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 981.369496] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 981.404279] env[61857]: INFO nova.scheduler.client.report [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted allocations for instance c3c9df36-d523-4a0f-9484-4446ad8c4138 [ 981.489279] env[61857]: DEBUG nova.network.neutron [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Updating instance_info_cache with network_info: [{"id": "7a4acc68-8312-44db-90f7-f628627266cc", "address": "fa:16:3e:79:ed:33", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a4acc68-83", "ovs_interfaceid": "7a4acc68-8312-44db-90f7-f628627266cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.696215] env[61857]: DEBUG oslo_concurrency.lockutils [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.696438] env[61857]: DEBUG oslo_concurrency.lockutils [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquired lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.696664] env[61857]: DEBUG nova.network.neutron [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 981.747009] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.747570] env[61857]: DEBUG nova.compute.manager [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 981.750139] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.663s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.755018] env[61857]: INFO nova.compute.claims [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.762621] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951409, 'name': ReconfigVM_Task, 'duration_secs': 0.180341} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.762884] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Reconfigured VM instance instance-00000035 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 981.763679] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b26cca1-2b59-4f4f-bb7d-9bf118f2211c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.786233] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 71920a62-090d-4df1-937e-87df3b043e28/71920a62-090d-4df1-937e-87df3b043e28.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 981.786572] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72b009e2-a87f-4a62-aa6c-381562ad1e03 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.811027] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 981.811027] env[61857]: value = "task-951410" [ 981.811027] env[61857]: _type = "Task" [ 981.811027] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.816904] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951410, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.909234] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.969040] env[61857]: DEBUG oslo_concurrency.lockutils [None req-15b68f0d-ad20-48df-b140-f435dc519811 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.281s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.991847] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "refresh_cache-2cd4f38e-4111-45b2-a8bd-dc0d38ada293" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.992247] env[61857]: DEBUG nova.compute.manager [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Instance network_info: |[{"id": "7a4acc68-8312-44db-90f7-f628627266cc", "address": "fa:16:3e:79:ed:33", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a4acc68-83", "ovs_interfaceid": "7a4acc68-8312-44db-90f7-f628627266cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 981.992591] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Acquired lock "refresh_cache-2cd4f38e-4111-45b2-a8bd-dc0d38ada293" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.992809] env[61857]: DEBUG nova.network.neutron [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Refreshing network info cache for port 7a4acc68-8312-44db-90f7-f628627266cc {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.994030] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:ed:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3a80436-f7a9-431a-acec-aca3d76e3f9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7a4acc68-8312-44db-90f7-f628627266cc', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.001774] env[61857]: DEBUG oslo.service.loopingcall [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.005062] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 982.005288] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c82eb26c-76e2-4eb2-9384-1d88003aceda {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.028847] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.028847] env[61857]: value = "task-951411" [ 982.028847] env[61857]: _type = "Task" [ 982.028847] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.042197] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951411, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.063517] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-c646ad3b-3087-4720-bac6-4c9b0108f02e" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.063827] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-c646ad3b-3087-4720-bac6-4c9b0108f02e" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.064328] env[61857]: DEBUG nova.objects.instance [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'flavor' on Instance uuid f982f0c5-c701-4e5b-b96d-c21a3c0da24f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.263531] env[61857]: DEBUG nova.compute.utils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 982.269491] env[61857]: DEBUG nova.compute.manager [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 982.269912] env[61857]: DEBUG nova.network.neutron [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 982.319671] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951410, 'name': ReconfigVM_Task, 'duration_secs': 0.270306} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.319973] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 71920a62-090d-4df1-937e-87df3b043e28/71920a62-090d-4df1-937e-87df3b043e28.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.320392] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance '71920a62-090d-4df1-937e-87df3b043e28' progress to 50 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 982.375294] env[61857]: DEBUG nova.policy [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2ef7e845007475a8a19cf3f520c3f51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1acf476aabc4166bc8505a3442367c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 982.389138] env[61857]: DEBUG nova.network.neutron [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Updated VIF entry in instance network info cache for port 7a4acc68-8312-44db-90f7-f628627266cc. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 982.389526] env[61857]: DEBUG nova.network.neutron [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Updating instance_info_cache with network_info: [{"id": "7a4acc68-8312-44db-90f7-f628627266cc", "address": "fa:16:3e:79:ed:33", "network": {"id": "2dcb03f4-1f25-4d3f-8f62-6e42a75865d9", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-902977930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bbcf7c2937040e1906e2273a07b671b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a4acc68-83", "ovs_interfaceid": "7a4acc68-8312-44db-90f7-f628627266cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.539503] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951411, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.568364] env[61857]: DEBUG nova.network.neutron [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updating instance_info_cache with network_info: [{"id": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "address": "fa:16:3e:dc:03:c9", "network": {"id": "d4dd4fc7-45cc-4dde-80aa-23506633d7ab", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1474003400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4725d163dec74a53940dc24ae88927f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5da8191c-5f", "ovs_interfaceid": "5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.738149] env[61857]: DEBUG nova.objects.instance [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'pci_requests' on Instance uuid f982f0c5-c701-4e5b-b96d-c21a3c0da24f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.769140] env[61857]: DEBUG nova.compute.manager [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 982.807689] env[61857]: DEBUG nova.network.neutron [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Successfully created port: c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 982.829042] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212e2545-ae81-40b1-9770-288c441984f3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.851443] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309daf8e-43ff-4a66-824b-e69daea8a7ad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.870608] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance '71920a62-090d-4df1-937e-87df3b043e28' progress to 67 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 982.893073] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Releasing lock "refresh_cache-2cd4f38e-4111-45b2-a8bd-dc0d38ada293" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.893358] env[61857]: DEBUG nova.compute.manager [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Received event network-vif-unplugged-5bd8b89d-fe07-48dc-a773-c222f7296a19 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.893561] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Acquiring lock "c3c9df36-d523-4a0f-9484-4446ad8c4138-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.893822] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.894058] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.894250] env[61857]: DEBUG nova.compute.manager [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] No waiting events found dispatching network-vif-unplugged-5bd8b89d-fe07-48dc-a773-c222f7296a19 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 982.894431] env[61857]: WARNING nova.compute.manager [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Received unexpected event network-vif-unplugged-5bd8b89d-fe07-48dc-a773-c222f7296a19 for instance with vm_state shelved and task_state shelving_offloading. [ 982.894690] env[61857]: DEBUG nova.compute.manager [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Received event network-changed-5bd8b89d-fe07-48dc-a773-c222f7296a19 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.895110] env[61857]: DEBUG nova.compute.manager [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Refreshing instance network info cache due to event network-changed-5bd8b89d-fe07-48dc-a773-c222f7296a19. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 982.895110] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Acquiring lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.895700] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Acquired lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.895700] env[61857]: DEBUG nova.network.neutron [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Refreshing network info cache for port 5bd8b89d-fe07-48dc-a773-c222f7296a19 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 982.971189] env[61857]: DEBUG oslo_concurrency.lockutils [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "c3c9df36-d523-4a0f-9484-4446ad8c4138" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.004756] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77d1877-6cd6-464c-9352-8e5c2d452a7d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.014342] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37e9680-011a-4ad4-99a5-8b4623174806 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.059893] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a533597-d460-4301-b7d0-670c3d308fe8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.068733] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951411, 'name': CreateVM_Task, 'duration_secs': 1.034949} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.070502] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 983.071344] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.071580] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.071982] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.073251] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0419d24a-78f8-4e35-885e-94224422ee1d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.077209] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9ce6a31-598b-4350-8f3a-88eacc355d9d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.079326] env[61857]: DEBUG oslo_concurrency.lockutils [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Releasing lock "refresh_cache-781a2790-e317-46fe-9be8-40c9e5f1f771" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.080214] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60e2987-0643-4f78-82f3-710323fd77ab {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.086170] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 983.086170] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52dcd305-7043-9993-e71a-65fed573a28c" [ 983.086170] env[61857]: _type = "Task" [ 983.086170] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.095653] env[61857]: DEBUG nova.compute.provider_tree [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.096930] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Resuming the VM {{(pid=61857) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 983.100168] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b05c72cc-468a-4b47-a719-08d257732a82 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.107613] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52dcd305-7043-9993-e71a-65fed573a28c, 'name': SearchDatastore_Task, 'duration_secs': 0.008976} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.109099] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.109357] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.109590] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.109739] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.109917] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.110260] env[61857]: DEBUG oslo_vmware.api [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 983.110260] env[61857]: value = "task-951412" [ 983.110260] env[61857]: _type = "Task" [ 983.110260] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.110999] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32068877-97cb-4eca-a6cd-b19717ccf56e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.120707] env[61857]: DEBUG oslo_vmware.api [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951412, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.122690] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.122877] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 983.123569] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf2f5147-2d32-4805-9d4a-2b9c751ce112 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.128293] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 983.128293] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52bf22ce-3d95-2e24-7f66-d813c6f185e0" [ 983.128293] env[61857]: _type = "Task" [ 983.128293] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.137319] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52bf22ce-3d95-2e24-7f66-d813c6f185e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.241297] env[61857]: DEBUG nova.objects.base [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 983.241586] env[61857]: DEBUG nova.network.neutron [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 983.313194] env[61857]: DEBUG nova.policy [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9056d344ff614950a60983450a2a40ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5653ed6ab0a948ef89bb03cf7426dd10', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 983.437131] env[61857]: DEBUG nova.network.neutron [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Port 90c3f890-dd3a-4d33-92c0-a6ec8632f14b binding to destination host cpu-1 is already ACTIVE {{(pid=61857) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 983.603809] env[61857]: DEBUG nova.scheduler.client.report [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.628422] env[61857]: DEBUG oslo_vmware.api [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951412, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.639061] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52bf22ce-3d95-2e24-7f66-d813c6f185e0, 'name': SearchDatastore_Task, 'duration_secs': 0.028969} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.639938] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48c93163-b3e5-4018-8956-682bd2c927df {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.646090] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 983.646090] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52cb637b-95fa-84f9-09b5-0fe11a2e3dd5" [ 983.646090] env[61857]: _type = "Task" [ 983.646090] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.655750] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52cb637b-95fa-84f9-09b5-0fe11a2e3dd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.664088] env[61857]: DEBUG nova.network.neutron [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Updated VIF entry in instance network info cache for port 5bd8b89d-fe07-48dc-a773-c222f7296a19. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 983.664479] env[61857]: DEBUG nova.network.neutron [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Updating instance_info_cache with network_info: [{"id": "5bd8b89d-fe07-48dc-a773-c222f7296a19", "address": "fa:16:3e:71:32:7d", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": null, "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap5bd8b89d-fe", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.778815] env[61857]: DEBUG nova.compute.manager [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.805168] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.805519] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.805827] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.806077] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.806242] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.806402] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.806662] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.806855] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.807053] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.807234] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.807418] env[61857]: DEBUG nova.virt.hardware [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.808810] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55be5ed4-9ef4-453a-92b2-15d93a56552c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.818532] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196a6526-af23-47e1-a43b-42eaa6031a9b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.104029] env[61857]: DEBUG oslo_concurrency.lockutils [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.104312] env[61857]: DEBUG oslo_concurrency.lockutils [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.104538] env[61857]: DEBUG oslo_concurrency.lockutils [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.104737] env[61857]: DEBUG oslo_concurrency.lockutils [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.104919] env[61857]: DEBUG oslo_concurrency.lockutils [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.107074] env[61857]: INFO nova.compute.manager [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Terminating instance [ 984.108787] env[61857]: DEBUG nova.compute.manager [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.108988] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 984.109849] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37902dcb-6831-4fda-822d-c6f17941c7fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.113155] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.113615] env[61857]: DEBUG nova.compute.manager [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 984.117135] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.207s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.117135] env[61857]: DEBUG nova.objects.instance [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lazy-loading 'resources' on Instance uuid c3c9df36-d523-4a0f-9484-4446ad8c4138 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.124673] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.125411] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93f1a4d7-a0a8-4632-b33c-f90e789d1c20 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.130637] env[61857]: DEBUG oslo_vmware.api [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951412, 'name': PowerOnVM_Task, 'duration_secs': 0.557051} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.131218] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Resumed the VM {{(pid=61857) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 984.131549] env[61857]: DEBUG nova.compute.manager [None req-707899fa-4435-461c-b604-d40aa560013c tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.132390] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a6b61d-cee1-4848-8ff2-fc889f0d58fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.135962] env[61857]: DEBUG oslo_vmware.api [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 984.135962] env[61857]: value = "task-951413" [ 984.135962] env[61857]: _type = "Task" [ 984.135962] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.156341] env[61857]: DEBUG oslo_vmware.api [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951413, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.161678] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52cb637b-95fa-84f9-09b5-0fe11a2e3dd5, 'name': SearchDatastore_Task, 'duration_secs': 0.024282} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.163616] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.163616] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 2cd4f38e-4111-45b2-a8bd-dc0d38ada293/2cd4f38e-4111-45b2-a8bd-dc0d38ada293.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 984.163616] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d995ac22-bddf-4628-ad67-d1ed3af07968 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.167020] env[61857]: DEBUG oslo_concurrency.lockutils [req-350f0df1-bf37-4c40-b881-e2e837737ed3 req-4e661add-b3e5-4712-bea8-4d50428ac4b4 service nova] Releasing lock "refresh_cache-c3c9df36-d523-4a0f-9484-4446ad8c4138" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.170606] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 984.170606] env[61857]: value = "task-951414" [ 984.170606] env[61857]: _type = "Task" [ 984.170606] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.179162] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951414, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.350389] env[61857]: DEBUG nova.network.neutron [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Successfully updated port: c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 984.458179] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "71920a62-090d-4df1-937e-87df3b043e28-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.458409] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.458566] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.621355] env[61857]: DEBUG nova.compute.utils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.623108] env[61857]: DEBUG nova.compute.manager [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 984.623108] env[61857]: DEBUG nova.network.neutron [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 984.625921] env[61857]: DEBUG nova.objects.instance [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lazy-loading 'numa_topology' on Instance uuid c3c9df36-d523-4a0f-9484-4446ad8c4138 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.654548] env[61857]: DEBUG oslo_vmware.api [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951413, 'name': PowerOffVM_Task, 'duration_secs': 0.429131} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.655753] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 984.655936] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 984.656992] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-88e98dc5-5ece-4f11-ab96-9a82e912e18a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.679777] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951414, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.694129] env[61857]: DEBUG nova.policy [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '324d98d433b94aeea5c489d47d705fe2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38d5d54ccbbe457a996d833b10868d2b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.724540] env[61857]: DEBUG nova.compute.manager [req-34cee33a-da0a-4f2f-a9c5-d6d9753552c6 req-401be37e-be95-4c54-98eb-bb80806ad41b service nova] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Received event network-vif-plugged-c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 984.724865] env[61857]: DEBUG oslo_concurrency.lockutils [req-34cee33a-da0a-4f2f-a9c5-d6d9753552c6 req-401be37e-be95-4c54-98eb-bb80806ad41b service nova] Acquiring lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.724993] env[61857]: DEBUG oslo_concurrency.lockutils [req-34cee33a-da0a-4f2f-a9c5-d6d9753552c6 req-401be37e-be95-4c54-98eb-bb80806ad41b service nova] Lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.725177] env[61857]: DEBUG oslo_concurrency.lockutils [req-34cee33a-da0a-4f2f-a9c5-d6d9753552c6 req-401be37e-be95-4c54-98eb-bb80806ad41b service nova] Lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.725359] env[61857]: DEBUG nova.compute.manager [req-34cee33a-da0a-4f2f-a9c5-d6d9753552c6 req-401be37e-be95-4c54-98eb-bb80806ad41b service nova] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] No waiting events found dispatching network-vif-plugged-c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 984.725522] env[61857]: WARNING nova.compute.manager [req-34cee33a-da0a-4f2f-a9c5-d6d9753552c6 req-401be37e-be95-4c54-98eb-bb80806ad41b service nova] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Received unexpected event network-vif-plugged-c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1 for instance with vm_state building and task_state spawning. [ 984.816303] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 984.816694] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 984.816959] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Deleting the datastore file [datastore1] d9d05629-79ef-45c6-ac54-b3bc18d306f4 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.817339] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ed5d98e-48ab-48ff-b6f2-65a8df59933b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.824338] env[61857]: DEBUG oslo_vmware.api [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 984.824338] env[61857]: value = "task-951416" [ 984.824338] env[61857]: _type = "Task" [ 984.824338] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.835702] env[61857]: DEBUG oslo_vmware.api [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951416, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.854944] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "refresh_cache-fb993549-1638-4aa5-bee3-9f303a7e5cfd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.855241] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "refresh_cache-fb993549-1638-4aa5-bee3-9f303a7e5cfd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.855489] env[61857]: DEBUG nova.network.neutron [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 985.101836] env[61857]: DEBUG nova.network.neutron [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Successfully created port: 77eb288b-3f38-448a-af1a-fada990f5ec4 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.126675] env[61857]: DEBUG nova.compute.manager [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 985.132377] env[61857]: DEBUG nova.objects.base [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 985.180117] env[61857]: DEBUG nova.compute.manager [req-20a9b7c0-49fa-4c64-b322-3997b1916fb9 req-cfa9c87c-7f6a-495d-ab84-6751f639349b service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received event network-vif-plugged-c646ad3b-3087-4720-bac6-4c9b0108f02e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 985.180523] env[61857]: DEBUG oslo_concurrency.lockutils [req-20a9b7c0-49fa-4c64-b322-3997b1916fb9 req-cfa9c87c-7f6a-495d-ab84-6751f639349b service nova] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.180807] env[61857]: DEBUG oslo_concurrency.lockutils [req-20a9b7c0-49fa-4c64-b322-3997b1916fb9 req-cfa9c87c-7f6a-495d-ab84-6751f639349b service nova] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.181013] env[61857]: DEBUG oslo_concurrency.lockutils [req-20a9b7c0-49fa-4c64-b322-3997b1916fb9 req-cfa9c87c-7f6a-495d-ab84-6751f639349b service nova] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.181298] env[61857]: DEBUG nova.compute.manager [req-20a9b7c0-49fa-4c64-b322-3997b1916fb9 req-cfa9c87c-7f6a-495d-ab84-6751f639349b service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] No waiting events found dispatching network-vif-plugged-c646ad3b-3087-4720-bac6-4c9b0108f02e {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 985.181752] env[61857]: WARNING nova.compute.manager [req-20a9b7c0-49fa-4c64-b322-3997b1916fb9 req-cfa9c87c-7f6a-495d-ab84-6751f639349b service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received unexpected event network-vif-plugged-c646ad3b-3087-4720-bac6-4c9b0108f02e for instance with vm_state active and task_state None. [ 985.185715] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951414, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.955018} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.188479] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 2cd4f38e-4111-45b2-a8bd-dc0d38ada293/2cd4f38e-4111-45b2-a8bd-dc0d38ada293.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 985.188741] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.189724] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0c9abeff-95b9-47be-bc26-c7517b896eed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.199059] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 985.199059] env[61857]: value = "task-951417" [ 985.199059] env[61857]: _type = "Task" [ 985.199059] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.213488] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951417, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.337882] env[61857]: DEBUG oslo_vmware.api [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951416, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.404662} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.339101] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.339101] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 985.339381] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 985.339381] env[61857]: INFO nova.compute.manager [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Took 1.23 seconds to destroy the instance on the hypervisor. [ 985.339637] env[61857]: DEBUG oslo.service.loopingcall [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.339850] env[61857]: DEBUG nova.compute.manager [-] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 985.339949] env[61857]: DEBUG nova.network.neutron [-] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 985.382230] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474ba3d5-0f98-48e6-83e6-00d5d33e2227 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.389534] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c019235-b0cd-4198-bbe2-195ab3510e47 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.447185] env[61857]: DEBUG nova.network.neutron [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 985.450944] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf1b079-915b-4e30-bc83-692994e1deb5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.465500] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35371bf9-06bc-47ba-a81c-6ba1998ea386 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.491216] env[61857]: DEBUG nova.compute.provider_tree [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 985.497322] env[61857]: DEBUG nova.network.neutron [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Successfully updated port: c646ad3b-3087-4720-bac6-4c9b0108f02e {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 985.579276] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.579377] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.579506] env[61857]: DEBUG nova.network.neutron [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 985.635287] env[61857]: DEBUG nova.network.neutron [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Updating instance_info_cache with network_info: [{"id": "c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1", "address": "fa:16:3e:15:11:20", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc89f8812-4f", "ovs_interfaceid": "c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.714439] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951417, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.271197} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.714732] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.715636] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9092187a-6986-4c79-a545-7aa60d42f77a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.741240] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 2cd4f38e-4111-45b2-a8bd-dc0d38ada293/2cd4f38e-4111-45b2-a8bd-dc0d38ada293.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.741620] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-764fb146-31c9-468a-a1a1-87180392dcc2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.764285] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 985.764285] env[61857]: value = "task-951418" [ 985.764285] env[61857]: _type = "Task" [ 985.764285] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.778906] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951418, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.003095] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.003357] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.003485] env[61857]: DEBUG nova.network.neutron [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 986.021902] env[61857]: ERROR nova.scheduler.client.report [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [req-bd804156-d0f0-4587-a092-131ffcc92ff8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2d5860d0-8574-4e55-9ced-20e66f0314c2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bd804156-d0f0-4587-a092-131ffcc92ff8"}]} [ 986.043036] env[61857]: DEBUG nova.scheduler.client.report [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Refreshing inventories for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 986.068140] env[61857]: DEBUG nova.scheduler.client.report [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Updating ProviderTree inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 986.068140] env[61857]: DEBUG nova.compute.provider_tree [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 40, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 986.084513] env[61857]: DEBUG nova.scheduler.client.report [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Refreshing aggregate associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, aggregates: None {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 986.129311] env[61857]: DEBUG nova.scheduler.client.report [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Refreshing trait associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 986.141248] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "refresh_cache-fb993549-1638-4aa5-bee3-9f303a7e5cfd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.141738] env[61857]: DEBUG nova.compute.manager [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Instance network_info: |[{"id": "c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1", "address": "fa:16:3e:15:11:20", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc89f8812-4f", "ovs_interfaceid": "c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 986.143830] env[61857]: DEBUG nova.compute.manager [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 986.147561] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:11:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.162407] env[61857]: DEBUG oslo.service.loopingcall [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.166280] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 986.168266] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d125d036-b2fb-4108-9bbd-2bd9f998e0fe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.193822] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.193822] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.193822] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.193822] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.193822] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.194141] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.194204] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.194381] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.194571] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.194765] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.194965] env[61857]: DEBUG nova.virt.hardware [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.195881] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b344de94-1483-465f-aa1d-f205602b0c40 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.200332] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 986.200332] env[61857]: value = "task-951419" [ 986.200332] env[61857]: _type = "Task" [ 986.200332] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.214755] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78bf4254-9712-4c01-815b-7bcdde1eacef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.222423] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951419, 'name': CreateVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.277315] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.382869] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae675ec-d142-46e0-8269-ce3746114cc4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.391350] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5bfc59-0f05-4237-9019-48c4ec8855d1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.396068] env[61857]: DEBUG nova.network.neutron [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance_info_cache with network_info: [{"id": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "address": "fa:16:3e:32:96:34", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c3f890-dd", "ovs_interfaceid": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.424531] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a955bb23-a8c5-4e58-ad51-55a1dc2322ae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.432728] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f08bf2-b03d-433f-ab7a-5255493790f0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.450983] env[61857]: DEBUG nova.compute.provider_tree [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 986.468131] env[61857]: DEBUG nova.compute.manager [req-3bd5a512-fa37-4dfd-bc86-8a965a34ca90 req-c84713d4-f3d3-4038-ac4f-7f431d1f62a5 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Received event network-vif-deleted-d9f80f94-9be0-4d62-863f-7562c6c9e63e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.468349] env[61857]: INFO nova.compute.manager [req-3bd5a512-fa37-4dfd-bc86-8a965a34ca90 req-c84713d4-f3d3-4038-ac4f-7f431d1f62a5 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Neutron deleted interface d9f80f94-9be0-4d62-863f-7562c6c9e63e; detaching it from the instance and deleting it from the info cache [ 986.468588] env[61857]: DEBUG nova.network.neutron [req-3bd5a512-fa37-4dfd-bc86-8a965a34ca90 req-c84713d4-f3d3-4038-ac4f-7f431d1f62a5 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.554321] env[61857]: WARNING nova.network.neutron [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] 30b7d27c-744c-46a6-8769-b1b1c95306e4 already exists in list: networks containing: ['30b7d27c-744c-46a6-8769-b1b1c95306e4']. ignoring it [ 986.554541] env[61857]: WARNING nova.network.neutron [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] 30b7d27c-744c-46a6-8769-b1b1c95306e4 already exists in list: networks containing: ['30b7d27c-744c-46a6-8769-b1b1c95306e4']. ignoring it [ 986.596529] env[61857]: DEBUG nova.network.neutron [-] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.711764] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951419, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.784891] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951418, 'name': ReconfigVM_Task, 'duration_secs': 0.596742} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.785331] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 2cd4f38e-4111-45b2-a8bd-dc0d38ada293/2cd4f38e-4111-45b2-a8bd-dc0d38ada293.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.786033] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66408a9b-25d6-483b-bc3e-31d8a6034b18 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.796745] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 986.796745] env[61857]: value = "task-951420" [ 986.796745] env[61857]: _type = "Task" [ 986.796745] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.808247] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951420, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.882058] env[61857]: DEBUG nova.compute.manager [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Received event network-changed-c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.882408] env[61857]: DEBUG nova.compute.manager [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Refreshing instance network info cache due to event network-changed-c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 986.882787] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] Acquiring lock "refresh_cache-fb993549-1638-4aa5-bee3-9f303a7e5cfd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.883055] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] Acquired lock "refresh_cache-fb993549-1638-4aa5-bee3-9f303a7e5cfd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.883340] env[61857]: DEBUG nova.network.neutron [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Refreshing network info cache for port c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 986.902021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.909452] env[61857]: DEBUG nova.network.neutron [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Successfully updated port: 77eb288b-3f38-448a-af1a-fada990f5ec4 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.971662] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-38e8bd0c-3b5a-4170-9c27-d2c6625f7ea9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.983057] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d110926a-5211-4187-9e7f-9b68c78d638b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.996194] env[61857]: DEBUG nova.scheduler.client.report [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Updated inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 986.997200] env[61857]: DEBUG nova.compute.provider_tree [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Updating resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 generation from 119 to 120 during operation: update_inventory {{(pid=61857) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 986.997200] env[61857]: DEBUG nova.compute.provider_tree [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 987.028009] env[61857]: DEBUG nova.compute.manager [req-3bd5a512-fa37-4dfd-bc86-8a965a34ca90 req-c84713d4-f3d3-4038-ac4f-7f431d1f62a5 service nova] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Detach interface failed, port_id=d9f80f94-9be0-4d62-863f-7562c6c9e63e, reason: Instance d9d05629-79ef-45c6-ac54-b3bc18d306f4 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 987.101312] env[61857]: INFO nova.compute.manager [-] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Took 1.76 seconds to deallocate network for instance. [ 987.186075] env[61857]: DEBUG nova.network.neutron [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d306c021-073e-45f7-9916-f3a6bc7a7dff", "address": "fa:16:3e:52:3f:1e", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd306c021-07", "ovs_interfaceid": "d306c021-073e-45f7-9916-f3a6bc7a7dff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c646ad3b-3087-4720-bac6-4c9b0108f02e", "address": "fa:16:3e:9f:2a:52", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc646ad3b-30", "ovs_interfaceid": "c646ad3b-3087-4720-bac6-4c9b0108f02e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.211534] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951419, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.307491] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951420, 'name': Rename_Task, 'duration_secs': 0.151155} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.307834] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 987.308109] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64f3650c-24ea-4415-944a-92a93542847b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.315512] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 987.315512] env[61857]: value = "task-951421" [ 987.315512] env[61857]: _type = "Task" [ 987.315512] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.324339] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951421, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.416492] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquiring lock "refresh_cache-e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.416644] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquired lock "refresh_cache-e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.416793] env[61857]: DEBUG nova.network.neutron [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 987.435173] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72fb46a-1a3c-471a-acd3-2fa6d0b71b43 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.456323] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ae8915-800c-42b6-a276-4724c3f005d7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.467890] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance '71920a62-090d-4df1-937e-87df3b043e28' progress to 83 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 987.504815] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.389s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.592275] env[61857]: DEBUG nova.network.neutron [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Updated VIF entry in instance network info cache for port c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 987.592764] env[61857]: DEBUG nova.network.neutron [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Updating instance_info_cache with network_info: [{"id": "c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1", "address": "fa:16:3e:15:11:20", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc89f8812-4f", "ovs_interfaceid": "c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.608263] env[61857]: DEBUG oslo_concurrency.lockutils [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.608584] env[61857]: DEBUG oslo_concurrency.lockutils [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.608990] env[61857]: DEBUG nova.objects.instance [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lazy-loading 'resources' on Instance uuid d9d05629-79ef-45c6-ac54-b3bc18d306f4 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.689649] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.690162] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.690282] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.691148] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8804c73-dc0b-4736-a718-fc79beb5c4ee {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.710667] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.710918] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.711104] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.711318] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.711552] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.711737] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.711953] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.712139] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.712315] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.712488] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.712670] env[61857]: DEBUG nova.virt.hardware [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.718930] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Reconfiguring VM to attach interface {{(pid=61857) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 987.722042] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67b89d90-927d-4786-8adb-71219d771372 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.742553] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951419, 'name': CreateVM_Task, 'duration_secs': 1.426057} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.743822] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 987.744215] env[61857]: DEBUG oslo_vmware.api [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 987.744215] env[61857]: value = "task-951422" [ 987.744215] env[61857]: _type = "Task" [ 987.744215] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.745139] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.745420] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.745808] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 987.746726] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee428e77-8b79-4a25-b136-4b1021a7d945 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.756427] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 987.756427] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fad5c4-d293-0720-d1e2-12298253cbaa" [ 987.756427] env[61857]: _type = "Task" [ 987.756427] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.760098] env[61857]: DEBUG oslo_vmware.api [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951422, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.769540] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fad5c4-d293-0720-d1e2-12298253cbaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.826381] env[61857]: DEBUG oslo_vmware.api [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951421, 'name': PowerOnVM_Task, 'duration_secs': 0.476195} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.826675] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 987.826886] env[61857]: INFO nova.compute.manager [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Took 8.27 seconds to spawn the instance on the hypervisor. [ 987.827084] env[61857]: DEBUG nova.compute.manager [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.827857] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bd76d4-4283-46ab-a39a-1a64277fad9e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.952411] env[61857]: DEBUG nova.network.neutron [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 987.974650] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 987.974957] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7293c236-36b9-4882-b113-91e8c4d924b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.983564] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 987.983564] env[61857]: value = "task-951423" [ 987.983564] env[61857]: _type = "Task" [ 987.983564] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.993676] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951423, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.012931] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c3e243cd-0ab3-405d-95f2-409b61c06b30 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.692s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.014064] env[61857]: DEBUG oslo_concurrency.lockutils [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.043s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.014359] env[61857]: DEBUG oslo_concurrency.lockutils [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "c3c9df36-d523-4a0f-9484-4446ad8c4138-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.014651] env[61857]: DEBUG oslo_concurrency.lockutils [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.015188] env[61857]: DEBUG oslo_concurrency.lockutils [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.019928] env[61857]: INFO nova.compute.manager [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Terminating instance [ 988.025603] env[61857]: DEBUG nova.compute.manager [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.025603] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 988.026246] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0469ed01-79bd-49b2-bc41-a0cbf968307e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.038106] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c329da3-e2c0-4847-a2b3-3689467f9b9c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.076846] env[61857]: WARNING nova.virt.vmwareapi.vmops [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c3c9df36-d523-4a0f-9484-4446ad8c4138 could not be found. [ 988.077121] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 988.077387] env[61857]: INFO nova.compute.manager [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Took 0.05 seconds to destroy the instance on the hypervisor. [ 988.078028] env[61857]: DEBUG oslo.service.loopingcall [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.078028] env[61857]: DEBUG nova.compute.manager [-] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 988.078226] env[61857]: DEBUG nova.network.neutron [-] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 988.095426] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] Releasing lock "refresh_cache-fb993549-1638-4aa5-bee3-9f303a7e5cfd" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.095762] env[61857]: DEBUG nova.compute.manager [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received event network-changed-c646ad3b-3087-4720-bac6-4c9b0108f02e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 988.095961] env[61857]: DEBUG nova.compute.manager [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Refreshing instance network info cache due to event network-changed-c646ad3b-3087-4720-bac6-4c9b0108f02e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 988.096226] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] Acquiring lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.096419] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] Acquired lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.096610] env[61857]: DEBUG nova.network.neutron [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Refreshing network info cache for port c646ad3b-3087-4720-bac6-4c9b0108f02e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 988.144447] env[61857]: DEBUG nova.network.neutron [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Updating instance_info_cache with network_info: [{"id": "77eb288b-3f38-448a-af1a-fada990f5ec4", "address": "fa:16:3e:35:16:13", "network": {"id": "4745fdb8-85cd-4cbc-923d-d6a713ba7698", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-846023146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d5d54ccbbe457a996d833b10868d2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b8137fc-f23d-49b1-b19c-3123a5588f34", "external-id": "nsx-vlan-transportzone-709", "segmentation_id": 709, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77eb288b-3f", "ovs_interfaceid": "77eb288b-3f38-448a-af1a-fada990f5ec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.259918] env[61857]: DEBUG oslo_vmware.api [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.271558] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fad5c4-d293-0720-d1e2-12298253cbaa, 'name': SearchDatastore_Task, 'duration_secs': 0.024458} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.272067] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.272439] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.272843] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.273285] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.273585] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.273939] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27c7b387-d877-4ea1-b06b-4fb819ccb998 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.288038] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.288038] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 988.288038] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-947fcbbf-d5a7-4417-b384-6272d70bfd26 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.299059] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 988.299059] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524f611d-8dc1-11c1-bdb0-298a2dd1f520" [ 988.299059] env[61857]: _type = "Task" [ 988.299059] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.305601] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6612d15e-0e0d-4666-9d81-27d1ae9a8d31 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.317857] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caad71c1-e728-4b37-8a0e-69466f66c75f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.322377] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524f611d-8dc1-11c1-bdb0-298a2dd1f520, 'name': SearchDatastore_Task, 'duration_secs': 0.010602} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.323928] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-392dfc6b-73dd-4ad8-bb08-ec08092bf7ff {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.364145] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d09cbe-b850-4c0b-affa-bbc296aee183 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.367531] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 988.367531] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52c0796f-fb74-1260-3f7a-18b06e9610e5" [ 988.367531] env[61857]: _type = "Task" [ 988.367531] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.368212] env[61857]: INFO nova.compute.manager [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Took 19.05 seconds to build instance. [ 988.377429] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81bd546-2906-449f-9b6a-71aeb4e1c9e2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.385376] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c0796f-fb74-1260-3f7a-18b06e9610e5, 'name': SearchDatastore_Task, 'duration_secs': 0.0152} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.386869] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.387298] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] fb993549-1638-4aa5-bee3-9f303a7e5cfd/fb993549-1638-4aa5-bee3-9f303a7e5cfd.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 988.387943] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4aa8a76a-cef0-4bf4-953e-60235d82fe5f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.399355] env[61857]: DEBUG nova.compute.provider_tree [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.407918] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 988.407918] env[61857]: value = "task-951424" [ 988.407918] env[61857]: _type = "Task" [ 988.407918] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.418522] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951424, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.499511] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951423, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.646882] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Releasing lock "refresh_cache-e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.647311] env[61857]: DEBUG nova.compute.manager [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Instance network_info: |[{"id": "77eb288b-3f38-448a-af1a-fada990f5ec4", "address": "fa:16:3e:35:16:13", "network": {"id": "4745fdb8-85cd-4cbc-923d-d6a713ba7698", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-846023146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d5d54ccbbe457a996d833b10868d2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b8137fc-f23d-49b1-b19c-3123a5588f34", "external-id": "nsx-vlan-transportzone-709", "segmentation_id": 709, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77eb288b-3f", "ovs_interfaceid": "77eb288b-3f38-448a-af1a-fada990f5ec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 988.648026] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:16:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b8137fc-f23d-49b1-b19c-3123a5588f34', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '77eb288b-3f38-448a-af1a-fada990f5ec4', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 988.656153] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Creating folder: Project (38d5d54ccbbe457a996d833b10868d2b). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 988.656399] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee19336a-ede0-4f1c-86c3-1eac9223e6d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.671302] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Created folder: Project (38d5d54ccbbe457a996d833b10868d2b) in parent group-v214027. [ 988.671574] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Creating folder: Instances. Parent ref: group-v214190. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 988.671661] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a7aad81-df9d-4617-95a3-b0e237e74196 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.683871] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Created folder: Instances in parent group-v214190. [ 988.684164] env[61857]: DEBUG oslo.service.loopingcall [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.684375] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 988.684601] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e889d196-6a07-4a6a-a531-4367fd6e7327 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.707963] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 988.707963] env[61857]: value = "task-951427" [ 988.707963] env[61857]: _type = "Task" [ 988.707963] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.716947] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951427, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.759562] env[61857]: DEBUG oslo_vmware.api [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951422, 'name': ReconfigVM_Task, 'duration_secs': 0.814711} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.760205] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.760465] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Reconfigured VM to attach interface {{(pid=61857) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 988.836409] env[61857]: DEBUG nova.network.neutron [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updated VIF entry in instance network info cache for port c646ad3b-3087-4720-bac6-4c9b0108f02e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 988.836928] env[61857]: DEBUG nova.network.neutron [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d306c021-073e-45f7-9916-f3a6bc7a7dff", "address": "fa:16:3e:52:3f:1e", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd306c021-07", "ovs_interfaceid": "d306c021-073e-45f7-9916-f3a6bc7a7dff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c646ad3b-3087-4720-bac6-4c9b0108f02e", "address": "fa:16:3e:9f:2a:52", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc646ad3b-30", "ovs_interfaceid": "c646ad3b-3087-4720-bac6-4c9b0108f02e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.877260] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ae629305-38b5-4049-9b29-d9fbaac0bafb tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.565s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.903491] env[61857]: DEBUG nova.scheduler.client.report [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.920854] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951424, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.996963] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951423, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.083603] env[61857]: DEBUG nova.network.neutron [-] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.220545] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951427, 'name': CreateVM_Task, 'duration_secs': 0.461559} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.220711] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 989.223766] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.223952] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.224323] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 989.227228] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a02d7b20-fe0b-49b2-8ee7-a58ff1d1311e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.231124] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 989.231124] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525d4604-2bc2-75f9-9cc0-b5e80892af46" [ 989.231124] env[61857]: _type = "Task" [ 989.231124] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.240422] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525d4604-2bc2-75f9-9cc0-b5e80892af46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.266187] env[61857]: DEBUG oslo_concurrency.lockutils [None req-cb45e6b2-041e-4c9e-9c34-bf680fd8ea94 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-c646ad3b-3087-4720-bac6-4c9b0108f02e" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.202s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.315993] env[61857]: DEBUG nova.compute.manager [req-6ab16536-c489-49f0-81c5-ee3bbddce7d5 req-4c9232cf-f431-48a9-8e3a-1d775fae5063 service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Received event network-changed-77eb288b-3f38-448a-af1a-fada990f5ec4 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.316280] env[61857]: DEBUG nova.compute.manager [req-6ab16536-c489-49f0-81c5-ee3bbddce7d5 req-4c9232cf-f431-48a9-8e3a-1d775fae5063 service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Refreshing instance network info cache due to event network-changed-77eb288b-3f38-448a-af1a-fada990f5ec4. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 989.316581] env[61857]: DEBUG oslo_concurrency.lockutils [req-6ab16536-c489-49f0-81c5-ee3bbddce7d5 req-4c9232cf-f431-48a9-8e3a-1d775fae5063 service nova] Acquiring lock "refresh_cache-e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.316698] env[61857]: DEBUG oslo_concurrency.lockutils [req-6ab16536-c489-49f0-81c5-ee3bbddce7d5 req-4c9232cf-f431-48a9-8e3a-1d775fae5063 service nova] Acquired lock "refresh_cache-e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.316875] env[61857]: DEBUG nova.network.neutron [req-6ab16536-c489-49f0-81c5-ee3bbddce7d5 req-4c9232cf-f431-48a9-8e3a-1d775fae5063 service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Refreshing network info cache for port 77eb288b-3f38-448a-af1a-fada990f5ec4 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.339436] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] Releasing lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.339793] env[61857]: DEBUG nova.compute.manager [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Received event network-vif-plugged-77eb288b-3f38-448a-af1a-fada990f5ec4 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.339997] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] Acquiring lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.340220] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] Lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.340467] env[61857]: DEBUG oslo_concurrency.lockutils [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] Lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.340647] env[61857]: DEBUG nova.compute.manager [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] No waiting events found dispatching network-vif-plugged-77eb288b-3f38-448a-af1a-fada990f5ec4 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.340825] env[61857]: WARNING nova.compute.manager [req-4f4c6d99-7931-467d-80a2-ab3b7065f613 req-c52e1891-431a-4ff4-ac22-f2dcef5cf42a service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Received unexpected event network-vif-plugged-77eb288b-3f38-448a-af1a-fada990f5ec4 for instance with vm_state building and task_state spawning. [ 989.411073] env[61857]: DEBUG oslo_concurrency.lockutils [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.802s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.424406] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951424, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659449} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.424695] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] fb993549-1638-4aa5-bee3-9f303a7e5cfd/fb993549-1638-4aa5-bee3-9f303a7e5cfd.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 989.424921] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 989.425195] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7dc4abb2-59fc-4719-b4aa-891306fb8ca3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.433882] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 989.433882] env[61857]: value = "task-951428" [ 989.433882] env[61857]: _type = "Task" [ 989.433882] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.438072] env[61857]: INFO nova.scheduler.client.report [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Deleted allocations for instance d9d05629-79ef-45c6-ac54-b3bc18d306f4 [ 989.447085] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951428, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.497279] env[61857]: DEBUG oslo_vmware.api [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951423, 'name': PowerOnVM_Task, 'duration_secs': 1.250868} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.497632] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 989.497860] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f397c7f8-184e-4a7f-bf0d-ba7d91e2009b tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance '71920a62-090d-4df1-937e-87df3b043e28' progress to 100 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 989.586438] env[61857]: INFO nova.compute.manager [-] [instance: c3c9df36-d523-4a0f-9484-4446ad8c4138] Took 1.51 seconds to deallocate network for instance. [ 989.742472] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525d4604-2bc2-75f9-9cc0-b5e80892af46, 'name': SearchDatastore_Task, 'duration_secs': 0.010526} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.742845] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.743124] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.743378] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.743536] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.743722] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.743978] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bbbcc7dd-221d-4e2e-9c99-5625599b0a86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.752929] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.753138] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 989.753815] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9728030-ca8e-40dd-8790-b7f275061159 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.758998] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 989.758998] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5274dc94-97ff-b3dd-993e-f22d652971b7" [ 989.758998] env[61857]: _type = "Task" [ 989.758998] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.766792] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5274dc94-97ff-b3dd-993e-f22d652971b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.945407] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951428, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094112} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.948080] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.950823] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85f7dcb-0f2a-450a-bd42-1861647c808b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.953850] env[61857]: DEBUG oslo_concurrency.lockutils [None req-61dfc516-49fa-4cbf-bb57-c8fc295c227c tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "d9d05629-79ef-45c6-ac54-b3bc18d306f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.850s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.975562] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] fb993549-1638-4aa5-bee3-9f303a7e5cfd/fb993549-1638-4aa5-bee3-9f303a7e5cfd.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.976594] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e112db05-3fa8-4c8d-9520-782ae7079c22 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.001629] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 990.001629] env[61857]: value = "task-951429" [ 990.001629] env[61857]: _type = "Task" [ 990.001629] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.014024] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951429, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.031067] env[61857]: DEBUG nova.network.neutron [req-6ab16536-c489-49f0-81c5-ee3bbddce7d5 req-4c9232cf-f431-48a9-8e3a-1d775fae5063 service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Updated VIF entry in instance network info cache for port 77eb288b-3f38-448a-af1a-fada990f5ec4. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 990.031669] env[61857]: DEBUG nova.network.neutron [req-6ab16536-c489-49f0-81c5-ee3bbddce7d5 req-4c9232cf-f431-48a9-8e3a-1d775fae5063 service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Updating instance_info_cache with network_info: [{"id": "77eb288b-3f38-448a-af1a-fada990f5ec4", "address": "fa:16:3e:35:16:13", "network": {"id": "4745fdb8-85cd-4cbc-923d-d6a713ba7698", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-846023146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d5d54ccbbe457a996d833b10868d2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b8137fc-f23d-49b1-b19c-3123a5588f34", "external-id": "nsx-vlan-transportzone-709", "segmentation_id": 709, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77eb288b-3f", "ovs_interfaceid": "77eb288b-3f38-448a-af1a-fada990f5ec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.273515] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5274dc94-97ff-b3dd-993e-f22d652971b7, 'name': SearchDatastore_Task, 'duration_secs': 0.00885} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.274023] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29a6eea7-89d9-4956-904a-6932833a38d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.281510] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 990.281510] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ec284d-791e-6edb-4040-71fa68602b9e" [ 990.281510] env[61857]: _type = "Task" [ 990.281510] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.289980] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ec284d-791e-6edb-4040-71fa68602b9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.517836] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951429, 'name': ReconfigVM_Task, 'duration_secs': 0.361614} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.517836] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Reconfigured VM instance instance-0000005e to attach disk [datastore2] fb993549-1638-4aa5-bee3-9f303a7e5cfd/fb993549-1638-4aa5-bee3-9f303a7e5cfd.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 990.518201] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c5010f0-e279-409e-92c9-e2774f4398af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.529236] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 990.529236] env[61857]: value = "task-951430" [ 990.529236] env[61857]: _type = "Task" [ 990.529236] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.535670] env[61857]: DEBUG oslo_concurrency.lockutils [req-6ab16536-c489-49f0-81c5-ee3bbddce7d5 req-4c9232cf-f431-48a9-8e3a-1d775fae5063 service nova] Releasing lock "refresh_cache-e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.539621] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951430, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.611392] env[61857]: DEBUG oslo_concurrency.lockutils [None req-49e77cbe-71fd-4b9b-9c0e-7b582b780c93 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "c3c9df36-d523-4a0f-9484-4446ad8c4138" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.597s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.738581] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.739290] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.739290] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.739428] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.739612] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.742132] env[61857]: INFO nova.compute.manager [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Terminating instance [ 990.744162] env[61857]: DEBUG nova.compute.manager [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 990.744387] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 990.745363] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25493d72-df19-46b5-a7f2-b1049fece057 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.754713] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 990.755018] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c22a426-0ec1-47fe-86a7-7f8f28a13631 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.763918] env[61857]: DEBUG oslo_vmware.api [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 990.763918] env[61857]: value = "task-951431" [ 990.763918] env[61857]: _type = "Task" [ 990.763918] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.779998] env[61857]: DEBUG oslo_vmware.api [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951431, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.795332] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ec284d-791e-6edb-4040-71fa68602b9e, 'name': SearchDatastore_Task, 'duration_secs': 0.026196} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.795665] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.795937] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c/e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 990.796725] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c28f205-b0be-43ef-a937-2d1f27a676d7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.805274] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 990.805274] env[61857]: value = "task-951432" [ 990.805274] env[61857]: _type = "Task" [ 990.805274] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.816822] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951432, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.030085] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-d306c021-073e-45f7-9916-f3a6bc7a7dff" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.030222] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-d306c021-073e-45f7-9916-f3a6bc7a7dff" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.044769] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951430, 'name': Rename_Task, 'duration_secs': 0.267914} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.045704] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 991.046059] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c348786d-5f5f-4232-90b9-a2bdb6eaa8de {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.057611] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 991.057611] env[61857]: value = "task-951434" [ 991.057611] env[61857]: _type = "Task" [ 991.057611] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.070824] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951434, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.275997] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "71920a62-090d-4df1-937e-87df3b043e28" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.275997] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.276729] env[61857]: DEBUG nova.compute.manager [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Going to confirm migration 3 {{(pid=61857) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 991.277571] env[61857]: DEBUG oslo_vmware.api [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951431, 'name': PowerOffVM_Task, 'duration_secs': 0.291925} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.277828] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.278011] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 991.278320] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b487685-8b11-4b67-9c67-7f55c4ba7861 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.317949] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951432, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50797} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.318325] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c/e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 991.318576] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.318816] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bbe8c044-d69f-4e47-812b-67e3f0811dc8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.327051] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 991.327051] env[61857]: value = "task-951436" [ 991.327051] env[61857]: _type = "Task" [ 991.327051] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.338799] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951436, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.362069] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 991.362459] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 991.362810] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleting the datastore file [datastore2] 2cd4f38e-4111-45b2-a8bd-dc0d38ada293 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.363155] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a816d2d3-b043-4ba5-a1a1-2b4af92b292f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.370918] env[61857]: DEBUG oslo_vmware.api [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for the task: (returnval){ [ 991.370918] env[61857]: value = "task-951437" [ 991.370918] env[61857]: _type = "Task" [ 991.370918] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.380940] env[61857]: DEBUG oslo_vmware.api [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951437, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.536879] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.537121] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.538069] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50460b9c-624b-4f1f-b48a-b04ccf2328b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.563278] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99c8926-1908-4058-8e84-bfcbdc818848 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.573522] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951434, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.597674] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Reconfiguring VM to detach interface {{(pid=61857) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 991.598275] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-757a1069-d0e5-40dc-9e73-f3421671dcfc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.618859] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 991.618859] env[61857]: value = "task-951438" [ 991.618859] env[61857]: _type = "Task" [ 991.618859] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.627880] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.816259] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.816508] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.816700] env[61857]: DEBUG nova.network.neutron [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.816898] env[61857]: DEBUG nova.objects.instance [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'info_cache' on Instance uuid 71920a62-090d-4df1-937e-87df3b043e28 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.839647] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951436, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073056} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.839996] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.840867] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70df462c-8c61-4f18-8dbf-8acde103d0b0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.865144] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c/e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.865394] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb4fe39b-a1af-455a-ad38-8053d7391fae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.888296] env[61857]: DEBUG oslo_vmware.api [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Task: {'id': task-951437, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130308} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.889495] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 991.889714] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 991.889903] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 991.890101] env[61857]: INFO nova.compute.manager [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Took 1.15 seconds to destroy the instance on the hypervisor. [ 991.890373] env[61857]: DEBUG oslo.service.loopingcall [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.890640] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 991.890640] env[61857]: value = "task-951439" [ 991.890640] env[61857]: _type = "Task" [ 991.890640] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.890833] env[61857]: DEBUG nova.compute.manager [-] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 991.890931] env[61857]: DEBUG nova.network.neutron [-] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 991.900486] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951439, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.909252] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.909492] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.069328] env[61857]: DEBUG oslo_vmware.api [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951434, 'name': PowerOnVM_Task, 'duration_secs': 0.56778} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.069636] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 992.069853] env[61857]: INFO nova.compute.manager [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Took 8.29 seconds to spawn the instance on the hypervisor. [ 992.070049] env[61857]: DEBUG nova.compute.manager [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.070848] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ee36d6-c1de-41f7-bf42-bc109b4348a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.129878] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.191298] env[61857]: DEBUG nova.compute.manager [req-78aff57d-9a16-4a52-8422-78f30e157ca2 req-fda87255-27ae-41d4-bc1c-c8f89e0bf5e4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Received event network-vif-deleted-7a4acc68-8312-44db-90f7-f628627266cc {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.191298] env[61857]: INFO nova.compute.manager [req-78aff57d-9a16-4a52-8422-78f30e157ca2 req-fda87255-27ae-41d4-bc1c-c8f89e0bf5e4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Neutron deleted interface 7a4acc68-8312-44db-90f7-f628627266cc; detaching it from the instance and deleting it from the info cache [ 992.191443] env[61857]: DEBUG nova.network.neutron [req-78aff57d-9a16-4a52-8422-78f30e157ca2 req-fda87255-27ae-41d4-bc1c-c8f89e0bf5e4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.403000] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951439, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.411652] env[61857]: DEBUG nova.compute.manager [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 992.591408] env[61857]: INFO nova.compute.manager [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Took 16.44 seconds to build instance. [ 992.631126] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.665986] env[61857]: DEBUG nova.network.neutron [-] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.694336] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9add3994-c0e4-4ebb-9c6c-cc4031a1f258 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.706115] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4afed4c-cbcd-483e-ba58-d2c1e64cc8bf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.746253] env[61857]: DEBUG nova.compute.manager [req-78aff57d-9a16-4a52-8422-78f30e157ca2 req-fda87255-27ae-41d4-bc1c-c8f89e0bf5e4 service nova] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Detach interface failed, port_id=7a4acc68-8312-44db-90f7-f628627266cc, reason: Instance 2cd4f38e-4111-45b2-a8bd-dc0d38ada293 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 992.904559] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951439, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.934884] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.935194] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.936667] env[61857]: INFO nova.compute.claims [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 992.959832] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "781a2790-e317-46fe-9be8-40c9e5f1f771" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.960156] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.960411] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.960613] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.960842] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.963165] env[61857]: INFO nova.compute.manager [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Terminating instance [ 992.965360] env[61857]: DEBUG nova.compute.manager [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 992.965593] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 992.966508] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a8fc6f-d1aa-4073-b071-d2f71cc17da9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.977938] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 992.977938] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-243568fe-527a-4f66-8113-5692248671b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.986512] env[61857]: DEBUG oslo_vmware.api [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 992.986512] env[61857]: value = "task-951440" [ 992.986512] env[61857]: _type = "Task" [ 992.986512] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.996399] env[61857]: DEBUG oslo_vmware.api [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951440, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.090154] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7b01eee6-c803-4d03-ab94-1078ace3769d tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.955s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.101017] env[61857]: DEBUG nova.network.neutron [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance_info_cache with network_info: [{"id": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "address": "fa:16:3e:32:96:34", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c3f890-dd", "ovs_interfaceid": "90c3f890-dd3a-4d33-92c0-a6ec8632f14b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.132055] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.168703] env[61857]: INFO nova.compute.manager [-] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Took 1.28 seconds to deallocate network for instance. [ 993.403419] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951439, 'name': ReconfigVM_Task, 'duration_secs': 1.169752} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.403762] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Reconfigured VM instance instance-0000005f to attach disk [datastore1] e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c/e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.404424] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2eb555f-3dc4-4b44-ae1d-82a0144b1672 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.411513] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 993.411513] env[61857]: value = "task-951441" [ 993.411513] env[61857]: _type = "Task" [ 993.411513] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.420602] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951441, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.496417] env[61857]: DEBUG oslo_vmware.api [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951440, 'name': PowerOffVM_Task, 'duration_secs': 0.212274} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.496698] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 993.496875] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 993.497175] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69e05a99-1fac-41b3-ab55-5613c6559ef3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.532240] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.532549] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.532856] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.533116] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.533304] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.535614] env[61857]: INFO nova.compute.manager [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Terminating instance [ 993.537444] env[61857]: DEBUG nova.compute.manager [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 993.537698] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 993.538636] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1f8050-3857-42e6-8844-a298f7183496 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.546650] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 993.546880] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd6aa7ca-384c-4abd-931b-355235053510 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.554538] env[61857]: DEBUG oslo_vmware.api [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 993.554538] env[61857]: value = "task-951443" [ 993.554538] env[61857]: _type = "Task" [ 993.554538] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.563300] env[61857]: DEBUG oslo_vmware.api [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951443, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.577089] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 993.577347] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 993.577542] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleting the datastore file [datastore1] 781a2790-e317-46fe-9be8-40c9e5f1f771 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 993.577818] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a5b1dbcf-c7d4-45f7-88ff-1e67e52b46f9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.584623] env[61857]: DEBUG oslo_vmware.api [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for the task: (returnval){ [ 993.584623] env[61857]: value = "task-951444" [ 993.584623] env[61857]: _type = "Task" [ 993.584623] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.593630] env[61857]: DEBUG oslo_vmware.api [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951444, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.604349] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-71920a62-090d-4df1-937e-87df3b043e28" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.604599] env[61857]: DEBUG nova.objects.instance [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'migration_context' on Instance uuid 71920a62-090d-4df1-937e-87df3b043e28 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.631417] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.675410] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.922836] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951441, 'name': Rename_Task, 'duration_secs': 0.170526} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.923179] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 993.923446] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26c71f6f-aeca-4335-82c7-5d62972c8a56 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.929960] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 993.929960] env[61857]: value = "task-951445" [ 993.929960] env[61857]: _type = "Task" [ 993.929960] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.938175] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951445, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.066177] env[61857]: DEBUG oslo_vmware.api [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951443, 'name': PowerOffVM_Task, 'duration_secs': 0.199396} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.068992] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 994.069203] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 994.069676] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-91847b9e-d33a-4a6e-b175-7a4b69bb208c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.097414] env[61857]: DEBUG oslo_vmware.api [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Task: {'id': task-951444, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138228} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.097561] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.097750] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 994.097932] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 994.098130] env[61857]: INFO nova.compute.manager [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Took 1.13 seconds to destroy the instance on the hypervisor. [ 994.098418] env[61857]: DEBUG oslo.service.loopingcall [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.098623] env[61857]: DEBUG nova.compute.manager [-] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 994.098718] env[61857]: DEBUG nova.network.neutron [-] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 994.109575] env[61857]: DEBUG nova.objects.base [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Object Instance<71920a62-090d-4df1-937e-87df3b043e28> lazy-loaded attributes: info_cache,migration_context {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 994.110724] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2650843a-fabf-4fa3-9807-4fefb1aca71e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.136960] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eaf69df-0da5-46df-89f8-ac0474a929e5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.146250] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.150340] env[61857]: DEBUG oslo_vmware.api [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 994.150340] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b958d9-a728-6ef9-56ea-e99618e75f01" [ 994.150340] env[61857]: _type = "Task" [ 994.150340] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.150649] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 994.150848] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 994.151040] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleting the datastore file [datastore2] fb993549-1638-4aa5-bee3-9f303a7e5cfd {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.152446] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-910e396e-2558-41e1-a805-9693d8842e37 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.154787] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77942a8e-1ef6-4943-bebb-fef2ed970e00 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.167267] env[61857]: DEBUG oslo_vmware.api [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b958d9-a728-6ef9-56ea-e99618e75f01, 'name': SearchDatastore_Task, 'duration_secs': 0.013147} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.169125] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662577df-ec60-4134-841e-51952a20d17a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.172267] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.172589] env[61857]: DEBUG oslo_vmware.api [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 994.172589] env[61857]: value = "task-951447" [ 994.172589] env[61857]: _type = "Task" [ 994.172589] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.200466] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea88ce6-fefd-44b8-9e3b-efd639d3f168 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.206073] env[61857]: DEBUG oslo_vmware.api [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951447, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.211464] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9555ce3-f1e8-4029-ab97-788306d1cba5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.225742] env[61857]: DEBUG nova.compute.provider_tree [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.424040] env[61857]: DEBUG nova.compute.manager [req-e8367da0-4dc5-4ac2-a2e2-f496bd2d22ec req-09294d21-fef8-4dfc-8887-a2bac3c79903 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Received event network-vif-deleted-5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 994.424040] env[61857]: INFO nova.compute.manager [req-e8367da0-4dc5-4ac2-a2e2-f496bd2d22ec req-09294d21-fef8-4dfc-8887-a2bac3c79903 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Neutron deleted interface 5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034; detaching it from the instance and deleting it from the info cache [ 994.424325] env[61857]: DEBUG nova.network.neutron [req-e8367da0-4dc5-4ac2-a2e2-f496bd2d22ec req-09294d21-fef8-4dfc-8887-a2bac3c79903 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.441282] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951445, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.641440] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.683849] env[61857]: DEBUG oslo_vmware.api [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951447, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.286042} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.684117] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.684319] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 994.684506] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 994.684696] env[61857]: INFO nova.compute.manager [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Took 1.15 seconds to destroy the instance on the hypervisor. [ 994.684931] env[61857]: DEBUG oslo.service.loopingcall [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.685153] env[61857]: DEBUG nova.compute.manager [-] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 994.685247] env[61857]: DEBUG nova.network.neutron [-] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 994.729885] env[61857]: DEBUG nova.scheduler.client.report [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.900798] env[61857]: DEBUG nova.network.neutron [-] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.926688] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bac07dfe-c178-488c-b836-68b03e1043bb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.938154] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8111e3-8dce-4d85-aaa3-6715cb798f29 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.951888] env[61857]: DEBUG oslo_vmware.api [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951445, 'name': PowerOnVM_Task, 'duration_secs': 0.74205} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.952628] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 994.952865] env[61857]: INFO nova.compute.manager [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Took 8.81 seconds to spawn the instance on the hypervisor. [ 994.953072] env[61857]: DEBUG nova.compute.manager [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 994.953801] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757284de-b532-4256-acd1-7aabf6f1c360 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.969603] env[61857]: DEBUG nova.compute.manager [req-e8367da0-4dc5-4ac2-a2e2-f496bd2d22ec req-09294d21-fef8-4dfc-8887-a2bac3c79903 service nova] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Detach interface failed, port_id=5da8191c-5fbf-4cbf-b1b8-a6ba1f0e2034, reason: Instance 781a2790-e317-46fe-9be8-40c9e5f1f771 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 995.145115] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.234599] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.235347] env[61857]: DEBUG nova.compute.manager [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 995.239205] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.564s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.239293] env[61857]: DEBUG nova.objects.instance [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lazy-loading 'resources' on Instance uuid 2cd4f38e-4111-45b2-a8bd-dc0d38ada293 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.377736] env[61857]: DEBUG nova.network.neutron [-] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.403351] env[61857]: INFO nova.compute.manager [-] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Took 1.30 seconds to deallocate network for instance. [ 995.468415] env[61857]: INFO nova.compute.manager [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Took 14.40 seconds to build instance. [ 995.642450] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.696175] env[61857]: DEBUG oslo_concurrency.lockutils [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "e91e99b9-1cd3-4345-af09-f14af4df1214" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.696468] env[61857]: DEBUG oslo_concurrency.lockutils [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "e91e99b9-1cd3-4345-af09-f14af4df1214" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.696686] env[61857]: DEBUG oslo_concurrency.lockutils [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "e91e99b9-1cd3-4345-af09-f14af4df1214-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.696883] env[61857]: DEBUG oslo_concurrency.lockutils [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "e91e99b9-1cd3-4345-af09-f14af4df1214-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.697140] env[61857]: DEBUG oslo_concurrency.lockutils [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "e91e99b9-1cd3-4345-af09-f14af4df1214-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.699633] env[61857]: INFO nova.compute.manager [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Terminating instance [ 995.701558] env[61857]: DEBUG nova.compute.manager [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 995.701769] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 995.702671] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d250c204-890a-47a1-813a-aa6015eb01e8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.710679] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 995.710953] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2da22172-9c84-4708-aefd-a1ee69d7881a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.717422] env[61857]: DEBUG oslo_vmware.api [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 995.717422] env[61857]: value = "task-951448" [ 995.717422] env[61857]: _type = "Task" [ 995.717422] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.725793] env[61857]: DEBUG oslo_vmware.api [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951448, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.745348] env[61857]: DEBUG nova.compute.utils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 995.746765] env[61857]: DEBUG nova.compute.manager [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 995.746928] env[61857]: DEBUG nova.network.neutron [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 995.786222] env[61857]: DEBUG nova.policy [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74f5499cbed541deaf1905605e10158f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62bd314daba846cd9ed3dc994dd1034b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 995.880459] env[61857]: INFO nova.compute.manager [-] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Took 1.20 seconds to deallocate network for instance. [ 995.912237] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.938842] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2678897b-9c95-4ee1-8a6e-c7a7a972510a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.948580] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5cf0cc4-acb7-4510-802a-99fb186e5be4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.984333] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7bd96146-a0b2-4951-a0d7-2e0f232e4825 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.927s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.987661] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d826ce-8c3c-4017-a011-684747071000 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.993777] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa8e181-73f3-4733-8e4e-c315b26df479 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.008850] env[61857]: DEBUG nova.compute.provider_tree [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.089718] env[61857]: DEBUG nova.network.neutron [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Successfully created port: 33d4e605-4d10-46de-8fe5-af0c42295576 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 996.144089] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.228899] env[61857]: DEBUG oslo_vmware.api [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951448, 'name': PowerOffVM_Task, 'duration_secs': 0.27289} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.229210] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 996.229388] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 996.229654] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-835da46a-54ca-44b7-a954-54b00d34726e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.250207] env[61857]: DEBUG nova.compute.manager [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 996.320640] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 996.320893] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 996.321116] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Deleting the datastore file [datastore2] e91e99b9-1cd3-4345-af09-f14af4df1214 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.321412] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18c107f1-195f-4d24-ac11-c6a2b05953b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.328872] env[61857]: DEBUG oslo_vmware.api [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 996.328872] env[61857]: value = "task-951450" [ 996.328872] env[61857]: _type = "Task" [ 996.328872] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.337601] env[61857]: DEBUG oslo_vmware.api [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951450, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.387424] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.470829] env[61857]: DEBUG nova.compute.manager [req-eae60a7d-4148-4ac7-b2f4-d0c55341d84a req-46b0357d-5388-4cf8-8af6-77bbce58fd8b service nova] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Received event network-vif-deleted-c89f8812-4f6e-4e0f-89ed-62b15f5a1ee1 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 996.511776] env[61857]: DEBUG nova.scheduler.client.report [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.644687] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.841317] env[61857]: DEBUG oslo_vmware.api [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951450, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224501} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.841642] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.841877] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 996.842107] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 996.842330] env[61857]: INFO nova.compute.manager [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Took 1.14 seconds to destroy the instance on the hypervisor. [ 996.842614] env[61857]: DEBUG oslo.service.loopingcall [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.842843] env[61857]: DEBUG nova.compute.manager [-] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 996.842963] env[61857]: DEBUG nova.network.neutron [-] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 997.017237] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.778s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.019805] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.847s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.045231] env[61857]: INFO nova.scheduler.client.report [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Deleted allocations for instance 2cd4f38e-4111-45b2-a8bd-dc0d38ada293 [ 997.144594] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.156422] env[61857]: DEBUG nova.compute.manager [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.157082] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d90b10b-8afe-4217-9a34-7777fb63442d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.260019] env[61857]: DEBUG nova.compute.manager [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 997.282992] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 997.283284] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 997.283476] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 997.283671] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 997.283828] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 997.283983] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 997.284216] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 997.284383] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 997.284559] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 997.284729] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 997.284911] env[61857]: DEBUG nova.virt.hardware [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 997.285854] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db37d36-075f-447b-99c3-c1ba614cbc6b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.295413] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a1a812-ae0e-4276-8d19-da1840c27c00 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.553171] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8de8a856-f583-431a-9101-92cf6fc5ebca tempest-ServerDiskConfigTestJSON-688122869 tempest-ServerDiskConfigTestJSON-688122869-project-member] Lock "2cd4f38e-4111-45b2-a8bd-dc0d38ada293" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.814s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.623385] env[61857]: DEBUG nova.network.neutron [-] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.625395] env[61857]: DEBUG nova.network.neutron [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Successfully updated port: 33d4e605-4d10-46de-8fe5-af0c42295576 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 997.645659] env[61857]: DEBUG oslo_vmware.api [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951438, 'name': ReconfigVM_Task, 'duration_secs': 5.979892} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.648649] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.648883] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Reconfigured VM to detach interface {{(pid=61857) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 997.669599] env[61857]: INFO nova.compute.manager [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] instance snapshotting [ 997.672245] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff19d280-9c50-495a-a4b7-e2b5924efcbd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.707113] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e22e47-73bd-4567-b68c-d8e970f5e038 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.737175] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312b4b99-500d-4a09-8b22-ebcb217bb579 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.745884] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7f30cd-3660-48e3-be75-a6b6ac7f8ba4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.780122] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701dd872-300e-4687-9501-455075c7bca3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.787823] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c17e0a19-0389-4fa8-aeb4-5ecde62853d3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.801546] env[61857]: DEBUG nova.compute.provider_tree [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.128126] env[61857]: INFO nova.compute.manager [-] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Took 1.28 seconds to deallocate network for instance. [ 998.128126] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "refresh_cache-94720bbd-a5db-45ca-9cfc-02c0f127c8bc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.128961] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "refresh_cache-94720bbd-a5db-45ca-9cfc-02c0f127c8bc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.128961] env[61857]: DEBUG nova.network.neutron [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 998.218384] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 998.218692] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e26d15bc-4ecd-433b-8fbe-62352015b174 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.229019] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 998.229019] env[61857]: value = "task-951451" [ 998.229019] env[61857]: _type = "Task" [ 998.229019] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.238610] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951451, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.305270] env[61857]: DEBUG nova.scheduler.client.report [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.504372] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.504755] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.507930] env[61857]: INFO nova.compute.manager [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Shelving [ 998.513423] env[61857]: DEBUG nova.compute.manager [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Received event network-vif-deleted-b2d821dc-4eb3-4cb4-962e-7b235495b9be {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 998.513509] env[61857]: DEBUG nova.compute.manager [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Received event network-vif-plugged-33d4e605-4d10-46de-8fe5-af0c42295576 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 998.513720] env[61857]: DEBUG oslo_concurrency.lockutils [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] Acquiring lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.513946] env[61857]: DEBUG oslo_concurrency.lockutils [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] Lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.514148] env[61857]: DEBUG oslo_concurrency.lockutils [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] Lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.514337] env[61857]: DEBUG nova.compute.manager [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] No waiting events found dispatching network-vif-plugged-33d4e605-4d10-46de-8fe5-af0c42295576 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 998.514529] env[61857]: WARNING nova.compute.manager [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Received unexpected event network-vif-plugged-33d4e605-4d10-46de-8fe5-af0c42295576 for instance with vm_state building and task_state spawning. [ 998.514706] env[61857]: DEBUG nova.compute.manager [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Received event network-changed-33d4e605-4d10-46de-8fe5-af0c42295576 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 998.514882] env[61857]: DEBUG nova.compute.manager [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Refreshing instance network info cache due to event network-changed-33d4e605-4d10-46de-8fe5-af0c42295576. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 998.515083] env[61857]: DEBUG oslo_concurrency.lockutils [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] Acquiring lock "refresh_cache-94720bbd-a5db-45ca-9cfc-02c0f127c8bc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.635185] env[61857]: DEBUG nova.compute.manager [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received event network-vif-deleted-c646ad3b-3087-4720-bac6-4c9b0108f02e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 998.635406] env[61857]: INFO nova.compute.manager [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Neutron deleted interface c646ad3b-3087-4720-bac6-4c9b0108f02e; detaching it from the instance and deleting it from the info cache [ 998.635647] env[61857]: DEBUG nova.network.neutron [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d306c021-073e-45f7-9916-f3a6bc7a7dff", "address": "fa:16:3e:52:3f:1e", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd306c021-07", "ovs_interfaceid": "d306c021-073e-45f7-9916-f3a6bc7a7dff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.637870] env[61857]: DEBUG oslo_concurrency.lockutils [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.665226] env[61857]: DEBUG nova.network.neutron [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 998.741177] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951451, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.829576] env[61857]: DEBUG nova.network.neutron [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Updating instance_info_cache with network_info: [{"id": "33d4e605-4d10-46de-8fe5-af0c42295576", "address": "fa:16:3e:00:cb:57", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33d4e605-4d", "ovs_interfaceid": "33d4e605-4d10-46de-8fe5-af0c42295576", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.990049] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.990049] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.990049] env[61857]: DEBUG nova.network.neutron [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.023196] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 999.023196] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9148c448-26b5-4b7f-99af-9a5f576c46ca {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.030455] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 999.030455] env[61857]: value = "task-951452" [ 999.030455] env[61857]: _type = "Task" [ 999.030455] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.042608] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.139883] env[61857]: DEBUG oslo_concurrency.lockutils [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.140125] env[61857]: DEBUG oslo_concurrency.lockutils [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Acquired lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.141110] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65dff2e3-3868-4d50-82b7-52f8dd82e8d0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.163339] env[61857]: DEBUG oslo_concurrency.lockutils [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.164248] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd1f8b2-fad9-4d2b-bb11-e149c1269554 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.192761] env[61857]: DEBUG nova.virt.vmwareapi.vmops [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Reconfiguring VM to detach interface {{(pid=61857) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 999.193181] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c77c653-cc7b-430d-aa24-e7c566f95cce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.213513] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Waiting for the task: (returnval){ [ 999.213513] env[61857]: value = "task-951453" [ 999.213513] env[61857]: _type = "Task" [ 999.213513] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.223131] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.243531] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951451, 'name': CreateSnapshot_Task, 'duration_secs': 0.625498} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.244251] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 999.244648] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317141f3-ba8a-4c54-a487-f3b389bdcc59 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.316008] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.296s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.319083] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.407s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.319442] env[61857]: DEBUG nova.objects.instance [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lazy-loading 'resources' on Instance uuid 781a2790-e317-46fe-9be8-40c9e5f1f771 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.332435] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "refresh_cache-94720bbd-a5db-45ca-9cfc-02c0f127c8bc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.334884] env[61857]: DEBUG nova.compute.manager [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Instance network_info: |[{"id": "33d4e605-4d10-46de-8fe5-af0c42295576", "address": "fa:16:3e:00:cb:57", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33d4e605-4d", "ovs_interfaceid": "33d4e605-4d10-46de-8fe5-af0c42295576", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 999.334884] env[61857]: DEBUG oslo_concurrency.lockutils [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] Acquired lock "refresh_cache-94720bbd-a5db-45ca-9cfc-02c0f127c8bc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.334884] env[61857]: DEBUG nova.network.neutron [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Refreshing network info cache for port 33d4e605-4d10-46de-8fe5-af0c42295576 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 999.339022] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:cb:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33d4e605-4d10-46de-8fe5-af0c42295576', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.344557] env[61857]: DEBUG oslo.service.loopingcall [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.346038] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 999.346038] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32bd2168-4fde-4a75-872b-52fa7db53adb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.366791] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 999.366791] env[61857]: value = "task-951454" [ 999.366791] env[61857]: _type = "Task" [ 999.366791] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.375482] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951454, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.541720] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951452, 'name': PowerOffVM_Task, 'duration_secs': 0.292412} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.542078] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 999.543146] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2af54a-0484-4bfe-a2d5-b16ceb6055c9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.565174] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36937773-c136-4687-80b8-1043f3aae1b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.728656] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.765220] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 999.765533] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3ab2a811-6b4d-4640-844e-c94c5b5b257e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.774882] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 999.774882] env[61857]: value = "task-951455" [ 999.774882] env[61857]: _type = "Task" [ 999.774882] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.783189] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951455, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.857738] env[61857]: INFO nova.network.neutron [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Port d306c021-073e-45f7-9916-f3a6bc7a7dff from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 999.857738] env[61857]: DEBUG nova.network.neutron [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.882667] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951454, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.893037] env[61857]: INFO nova.scheduler.client.report [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleted allocation for migration 29cccd98-cdb2-46c7-920c-198017f1584b [ 1000.047170] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6e18ce-73f8-4174-a390-f5ad9c008f5a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.057637] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61c774f-f855-4be3-bec1-54b351efdb0e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.100425] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1000.101681] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6110f771-4931-4cf9-bcde-8970d1657117 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.104144] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2383efaa-55e8-4305-a0f6-956a3b5edbfa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.121022] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e742033f-ac2a-42e3-8cbf-734c86b92f7a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.124122] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1000.124122] env[61857]: value = "task-951456" [ 1000.124122] env[61857]: _type = "Task" [ 1000.124122] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.140483] env[61857]: DEBUG nova.compute.provider_tree [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.148106] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951456, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.225231] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.281929] env[61857]: DEBUG nova.network.neutron [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Updated VIF entry in instance network info cache for port 33d4e605-4d10-46de-8fe5-af0c42295576. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1000.282293] env[61857]: DEBUG nova.network.neutron [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Updating instance_info_cache with network_info: [{"id": "33d4e605-4d10-46de-8fe5-af0c42295576", "address": "fa:16:3e:00:cb:57", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33d4e605-4d", "ovs_interfaceid": "33d4e605-4d10-46de-8fe5-af0c42295576", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.287288] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951455, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.359280] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.380305] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951454, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.404045] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f7b6d755-c66d-4c69-9190-ca5a188e8332 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.126s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.636248] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951456, 'name': CreateSnapshot_Task, 'duration_secs': 0.485504} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.636958] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1000.637475] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda26189-d90f-49df-a3bc-eaa75854fc80 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.644114] env[61857]: DEBUG nova.scheduler.client.report [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.728545] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.786546] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951455, 'name': CloneVM_Task} progress is 95%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.788189] env[61857]: DEBUG oslo_concurrency.lockutils [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] Releasing lock "refresh_cache-94720bbd-a5db-45ca-9cfc-02c0f127c8bc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.788448] env[61857]: DEBUG nova.compute.manager [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received event network-vif-deleted-d306c021-073e-45f7-9916-f3a6bc7a7dff {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1000.788634] env[61857]: INFO nova.compute.manager [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Neutron deleted interface d306c021-073e-45f7-9916-f3a6bc7a7dff; detaching it from the instance and deleting it from the info cache [ 1000.789030] env[61857]: DEBUG nova.network.neutron [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [{"id": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "address": "fa:16:3e:16:00:c9", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d3974fc-a3", "ovs_interfaceid": "9d3974fc-a398-4220-9f8c-d78aa99796bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c646ad3b-3087-4720-bac6-4c9b0108f02e", "address": "fa:16:3e:9f:2a:52", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc646ad3b-30", "ovs_interfaceid": "c646ad3b-3087-4720-bac6-4c9b0108f02e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.864234] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1e8a280f-fdff-4dc8-8ed4-1378ab62b4e8 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-f982f0c5-c701-4e5b-b96d-c21a3c0da24f-d306c021-073e-45f7-9916-f3a6bc7a7dff" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.834s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.882218] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951454, 'name': CreateVM_Task, 'duration_secs': 1.412283} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.882404] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1000.883247] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.883356] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.883689] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1000.883947] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff0debd2-6927-49be-9751-1c2f2dbf4cc6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.889560] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1000.889560] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521bbfae-9932-df68-04a2-27681718317f" [ 1000.889560] env[61857]: _type = "Task" [ 1000.889560] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.899637] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521bbfae-9932-df68-04a2-27681718317f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.938539] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "71920a62-090d-4df1-937e-87df3b043e28" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.938814] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.939041] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "71920a62-090d-4df1-937e-87df3b043e28-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.939239] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.939417] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.941718] env[61857]: INFO nova.compute.manager [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Terminating instance [ 1000.944158] env[61857]: DEBUG nova.compute.manager [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.944389] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1000.945326] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b40e20b-5db6-408c-af02-b1149087d8b4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.953854] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1000.954364] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e389d43-852f-4038-93ee-d3aa3de15310 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.961547] env[61857]: DEBUG oslo_vmware.api [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1000.961547] env[61857]: value = "task-951457" [ 1000.961547] env[61857]: _type = "Task" [ 1000.961547] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.970476] env[61857]: DEBUG oslo_vmware.api [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951457, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.155570] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.836s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.164191] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1001.164766] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.778s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.164997] env[61857]: DEBUG nova.objects.instance [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lazy-loading 'resources' on Instance uuid fb993549-1638-4aa5-bee3-9f303a7e5cfd {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.166017] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1aba3ebe-550b-4cd2-91f1-58b2903e5224 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.175683] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1001.175683] env[61857]: value = "task-951458" [ 1001.175683] env[61857]: _type = "Task" [ 1001.175683] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.184778] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951458, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.187759] env[61857]: INFO nova.scheduler.client.report [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Deleted allocations for instance 781a2790-e317-46fe-9be8-40c9e5f1f771 [ 1001.226235] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.289301] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951455, 'name': CloneVM_Task, 'duration_secs': 1.312373} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.289641] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Created linked-clone VM from snapshot [ 1001.290521] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba7ba59-cd4d-43bb-aeea-edd5fb466eb2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.297548] env[61857]: DEBUG oslo_concurrency.lockutils [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.303899] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Uploading image 9ed0a868-4498-44da-a97c-65b2330033f4 {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1001.340609] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1001.340609] env[61857]: value = "vm-214195" [ 1001.340609] env[61857]: _type = "VirtualMachine" [ 1001.340609] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1001.340609] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-799a32da-8ff0-408f-849c-6bf5e97339cf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.346728] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lease: (returnval){ [ 1001.346728] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fe20ff-bc8e-7b58-dabb-295955cf669b" [ 1001.346728] env[61857]: _type = "HttpNfcLease" [ 1001.346728] env[61857]: } obtained for exporting VM: (result){ [ 1001.346728] env[61857]: value = "vm-214195" [ 1001.346728] env[61857]: _type = "VirtualMachine" [ 1001.346728] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1001.346981] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the lease: (returnval){ [ 1001.346981] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fe20ff-bc8e-7b58-dabb-295955cf669b" [ 1001.346981] env[61857]: _type = "HttpNfcLease" [ 1001.346981] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1001.354993] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1001.354993] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fe20ff-bc8e-7b58-dabb-295955cf669b" [ 1001.354993] env[61857]: _type = "HttpNfcLease" [ 1001.354993] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1001.400968] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521bbfae-9932-df68-04a2-27681718317f, 'name': SearchDatastore_Task, 'duration_secs': 0.010729} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.401386] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.403230] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1001.403230] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.403230] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.403230] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1001.403230] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5c7bfc8-eb68-433e-b9ee-606da55060ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.427902] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1001.428373] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1001.429061] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7425ca0-adbb-467e-a02f-f6214dff5fe4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.436026] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1001.436026] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521bbf14-4427-b57a-3f36-bfb0455a38e1" [ 1001.436026] env[61857]: _type = "Task" [ 1001.436026] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.444447] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521bbf14-4427-b57a-3f36-bfb0455a38e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.472064] env[61857]: DEBUG oslo_vmware.api [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951457, 'name': PowerOffVM_Task, 'duration_secs': 0.192554} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.472349] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1001.472528] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1001.472822] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb6050fb-1356-4fc4-961b-7dd7b5fa94ed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.559450] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1001.559450] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1001.559716] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleting the datastore file [datastore2] 71920a62-090d-4df1-937e-87df3b043e28 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.560046] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edfe2d19-f72c-457d-98cc-470f7ca69c0e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.568243] env[61857]: DEBUG oslo_vmware.api [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1001.568243] env[61857]: value = "task-951461" [ 1001.568243] env[61857]: _type = "Task" [ 1001.568243] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.576761] env[61857]: DEBUG oslo_vmware.api [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951461, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.688895] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951458, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.698959] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6c5de9ae-2118-4583-a5e5-48e8430b5d0a tempest-ServersNegativeTestJSON-1440553386 tempest-ServersNegativeTestJSON-1440553386-project-member] Lock "781a2790-e317-46fe-9be8-40c9e5f1f771" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.739s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.726265] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.833473] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137c4fa4-e9fe-4b07-85d6-6ecb29129136 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.843692] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e256737-1277-4b67-b43f-a7b93e192e15 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.854644] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1001.854644] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fe20ff-bc8e-7b58-dabb-295955cf669b" [ 1001.854644] env[61857]: _type = "HttpNfcLease" [ 1001.854644] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1001.878873] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1001.878873] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fe20ff-bc8e-7b58-dabb-295955cf669b" [ 1001.878873] env[61857]: _type = "HttpNfcLease" [ 1001.878873] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1001.880581] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574020f7-f547-439c-948e-7d0e3b296c32 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.883748] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd5bccc-8b1d-4a90-8046-d1f142801a5c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.895196] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65047347-6dcd-461a-b653-980b1ae60bd5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.899130] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdb239-3838-2467-9cb9-429242cbd8e6/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1001.899338] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdb239-3838-2467-9cb9-429242cbd8e6/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1001.968013] env[61857]: DEBUG nova.compute.provider_tree [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.977375] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521bbf14-4427-b57a-3f36-bfb0455a38e1, 'name': SearchDatastore_Task, 'duration_secs': 0.012555} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.982024] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-156a0705-e635-4814-806b-a450f2e2370f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.985687] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1001.985687] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ed1383-862d-4be8-bf67-3b1e1a3aed17" [ 1001.985687] env[61857]: _type = "Task" [ 1001.985687] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.995223] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ed1383-862d-4be8-bf67-3b1e1a3aed17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.019373] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cbc97301-c433-4aca-8246-8e6ee4a36017 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.079271] env[61857]: DEBUG oslo_vmware.api [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951461, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.278615} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.079608] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.079699] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1002.079885] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1002.080084] env[61857]: INFO nova.compute.manager [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1002.080383] env[61857]: DEBUG oslo.service.loopingcall [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.080595] env[61857]: DEBUG nova.compute.manager [-] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1002.080690] env[61857]: DEBUG nova.network.neutron [-] [instance: 71920a62-090d-4df1-937e-87df3b043e28] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1002.188349] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951458, 'name': CloneVM_Task} progress is 95%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.226830] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.473936] env[61857]: DEBUG nova.scheduler.client.report [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.502927] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ed1383-862d-4be8-bf67-3b1e1a3aed17, 'name': SearchDatastore_Task, 'duration_secs': 0.009765} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.502927] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.502927] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 94720bbd-a5db-45ca-9cfc-02c0f127c8bc/94720bbd-a5db-45ca-9cfc-02c0f127c8bc.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1002.502927] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-825f4de5-e58d-496a-bed5-d93d7c4e81ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.513287] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1002.513287] env[61857]: value = "task-951462" [ 1002.513287] env[61857]: _type = "Task" [ 1002.513287] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.522928] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951462, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.695304] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951458, 'name': CloneVM_Task, 'duration_secs': 1.334671} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.696553] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Created linked-clone VM from snapshot [ 1002.697527] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c576c1-19a0-483f-a918-758656ac69d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.710476] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Uploading image 6f2ff642-ffbc-477c-83d9-8ce6daf6b818 {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1002.741296] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.745052] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1002.745052] env[61857]: value = "vm-214197" [ 1002.745052] env[61857]: _type = "VirtualMachine" [ 1002.745052] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1002.746232] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f553f985-c322-4aeb-9793-a1c34689f3ce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.757943] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lease: (returnval){ [ 1002.757943] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524b3e7b-a58e-c9c6-76cd-0e572f9b4dfd" [ 1002.757943] env[61857]: _type = "HttpNfcLease" [ 1002.757943] env[61857]: } obtained for exporting VM: (result){ [ 1002.757943] env[61857]: value = "vm-214197" [ 1002.757943] env[61857]: _type = "VirtualMachine" [ 1002.757943] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1002.758581] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the lease: (returnval){ [ 1002.758581] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524b3e7b-a58e-c9c6-76cd-0e572f9b4dfd" [ 1002.758581] env[61857]: _type = "HttpNfcLease" [ 1002.758581] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1002.769354] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1002.769354] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524b3e7b-a58e-c9c6-76cd-0e572f9b4dfd" [ 1002.769354] env[61857]: _type = "HttpNfcLease" [ 1002.769354] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1002.984743] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.820s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.991485] env[61857]: DEBUG oslo_concurrency.lockutils [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.353s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.992590] env[61857]: DEBUG nova.objects.instance [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lazy-loading 'resources' on Instance uuid e91e99b9-1cd3-4345-af09-f14af4df1214 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.023382] env[61857]: INFO nova.scheduler.client.report [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted allocations for instance fb993549-1638-4aa5-bee3-9f303a7e5cfd [ 1003.038713] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951462, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.231649] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.270022] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1003.270022] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524b3e7b-a58e-c9c6-76cd-0e572f9b4dfd" [ 1003.270022] env[61857]: _type = "HttpNfcLease" [ 1003.270022] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1003.270022] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1003.270022] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524b3e7b-a58e-c9c6-76cd-0e572f9b4dfd" [ 1003.270022] env[61857]: _type = "HttpNfcLease" [ 1003.270022] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1003.270022] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f8a94c-c4a5-4b21-b66f-98da7ad29bcf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.291152] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217acb-3c7c-c3f0-293a-db11752dadc7/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1003.291152] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217acb-3c7c-c3f0-293a-db11752dadc7/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1003.296384] env[61857]: DEBUG nova.compute.manager [req-a105026e-893b-467e-9c0b-39562361b8ef req-41958bf1-0526-48f4-9a69-49567c7daf95 service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Received event network-vif-deleted-90c3f890-dd3a-4d33-92c0-a6ec8632f14b {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1003.296384] env[61857]: INFO nova.compute.manager [req-a105026e-893b-467e-9c0b-39562361b8ef req-41958bf1-0526-48f4-9a69-49567c7daf95 service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Neutron deleted interface 90c3f890-dd3a-4d33-92c0-a6ec8632f14b; detaching it from the instance and deleting it from the info cache [ 1003.296384] env[61857]: DEBUG nova.network.neutron [req-a105026e-893b-467e-9c0b-39562361b8ef req-41958bf1-0526-48f4-9a69-49567c7daf95 service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.419600] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8d9d0847-3294-4d43-a7a4-b32e6ef38152 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.524614] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951462, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.544413] env[61857]: DEBUG oslo_concurrency.lockutils [None req-4dfa2cb2-a9da-42ba-b7fa-1ccd484319c5 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "fb993549-1638-4aa5-bee3-9f303a7e5cfd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.012s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.696366] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10bb6136-68c7-4d03-aa78-07537e010035 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.711422] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc0a407-9cfd-42eb-9c9f-a7c69f581f69 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.746667] env[61857]: DEBUG nova.network.neutron [-] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.753316] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779eda45-cfdc-4607-bb4c-763822a6cd37 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.761733] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.766645] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c6b4fd-3ede-46f8-94b9-be06cd0634aa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.789895] env[61857]: DEBUG nova.compute.provider_tree [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.799424] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-251bba55-6b32-4a6f-a462-9b0ba89530c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.810641] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f441a62a-631a-4cd1-81a6-195207081501 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.846352] env[61857]: DEBUG nova.compute.manager [req-a105026e-893b-467e-9c0b-39562361b8ef req-41958bf1-0526-48f4-9a69-49567c7daf95 service nova] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Detach interface failed, port_id=90c3f890-dd3a-4d33-92c0-a6ec8632f14b, reason: Instance 71920a62-090d-4df1-937e-87df3b043e28 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1004.027232] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951462, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.508146} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.028071] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 94720bbd-a5db-45ca-9cfc-02c0f127c8bc/94720bbd-a5db-45ca-9cfc-02c0f127c8bc.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1004.028184] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.028533] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1c52ec4-a96d-4c31-916a-25a2d5108445 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.039329] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1004.039329] env[61857]: value = "task-951464" [ 1004.039329] env[61857]: _type = "Task" [ 1004.039329] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.053916] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951464, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.256594] env[61857]: INFO nova.compute.manager [-] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Took 2.17 seconds to deallocate network for instance. [ 1004.263821] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.297238] env[61857]: DEBUG nova.scheduler.client.report [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.551497] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951464, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099878} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.551959] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.552841] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0bc7dd-d559-4f05-a310-ef04b424ffcc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.579731] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 94720bbd-a5db-45ca-9cfc-02c0f127c8bc/94720bbd-a5db-45ca-9cfc-02c0f127c8bc.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.580347] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a285e810-63ac-419e-9b2f-9e6b20b894ef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.606276] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1004.606276] env[61857]: value = "task-951465" [ 1004.606276] env[61857]: _type = "Task" [ 1004.606276] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.616615] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951465, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.741188] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "ce67a15a-8604-4523-a8fa-a34ccf9914da" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.741562] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "ce67a15a-8604-4523-a8fa-a34ccf9914da" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.742964] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "ce67a15a-8604-4523-a8fa-a34ccf9914da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.743215] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "ce67a15a-8604-4523-a8fa-a34ccf9914da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.743435] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "ce67a15a-8604-4523-a8fa-a34ccf9914da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.747247] env[61857]: INFO nova.compute.manager [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Terminating instance [ 1004.752855] env[61857]: DEBUG nova.compute.manager [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1004.753079] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1004.753935] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e259cf8-2d03-4c94-b3fe-1777f06050ed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.765012] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.767741] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.768094] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2364eb3c-6b8d-46d9-ac63-9d99c35d085a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.773062] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.776923] env[61857]: DEBUG oslo_vmware.api [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1004.776923] env[61857]: value = "task-951466" [ 1004.776923] env[61857]: _type = "Task" [ 1004.776923] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.787078] env[61857]: DEBUG oslo_vmware.api [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951466, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.803228] env[61857]: DEBUG oslo_concurrency.lockutils [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.812s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.806502] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.033s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.806797] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.837688] env[61857]: INFO nova.scheduler.client.report [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Deleted allocations for instance e91e99b9-1cd3-4345-af09-f14af4df1214 [ 1004.841658] env[61857]: INFO nova.scheduler.client.report [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleted allocations for instance 71920a62-090d-4df1-937e-87df3b043e28 [ 1005.117740] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951465, 'name': ReconfigVM_Task, 'duration_secs': 0.349616} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.118315] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 94720bbd-a5db-45ca-9cfc-02c0f127c8bc/94720bbd-a5db-45ca-9cfc-02c0f127c8bc.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.119337] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e094a0b-c3e2-47ab-89a6-33d35dc61ec5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.128152] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1005.128152] env[61857]: value = "task-951467" [ 1005.128152] env[61857]: _type = "Task" [ 1005.128152] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.141323] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951467, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.259970] env[61857]: DEBUG oslo_vmware.api [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Task: {'id': task-951453, 'name': ReconfigVM_Task, 'duration_secs': 5.822076} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.260437] env[61857]: DEBUG oslo_concurrency.lockutils [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] Releasing lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.260712] env[61857]: DEBUG nova.virt.vmwareapi.vmops [req-df89d471-7fde-432b-bc2f-ec170305b64e req-36abf221-9f36-4cee-8712-0febcc90433f service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Reconfigured VM to detach interface {{(pid=61857) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1005.261484] env[61857]: DEBUG oslo_concurrency.lockutils [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.098s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.261784] env[61857]: DEBUG oslo_concurrency.lockutils [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.262119] env[61857]: DEBUG oslo_concurrency.lockutils [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.262416] env[61857]: DEBUG oslo_concurrency.lockutils [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.264768] env[61857]: INFO nova.compute.manager [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Terminating instance [ 1005.266824] env[61857]: DEBUG nova.compute.manager [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1005.267171] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1005.268170] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8de0c7-a535-4d2a-a54b-86e93d91323b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.278601] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1005.282797] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01537fbd-1b13-48c7-be68-ddaed18651bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.291283] env[61857]: DEBUG oslo_vmware.api [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951466, 'name': PowerOffVM_Task, 'duration_secs': 0.221931} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.292939] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1005.293224] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1005.293627] env[61857]: DEBUG oslo_vmware.api [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1005.293627] env[61857]: value = "task-951468" [ 1005.293627] env[61857]: _type = "Task" [ 1005.293627] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.293998] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e620635-82ee-4a41-a9f0-355d1365a11f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.306204] env[61857]: DEBUG oslo_vmware.api [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951468, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.355543] env[61857]: DEBUG oslo_concurrency.lockutils [None req-736f7f62-1b3d-4575-bfc6-e218cd45a272 tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "e91e99b9-1cd3-4345-af09-f14af4df1214" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.658s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.355543] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a39dd002-b7d3-494a-93eb-3991258ac7fa tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "71920a62-090d-4df1-937e-87df3b043e28" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.417s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.400172] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1005.400172] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1005.400172] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleting the datastore file [datastore1] ce67a15a-8604-4523-a8fa-a34ccf9914da {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.400839] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c40eaf0e-1f11-4da6-8ec3-829d61f7b088 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.410996] env[61857]: DEBUG oslo_vmware.api [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1005.410996] env[61857]: value = "task-951470" [ 1005.410996] env[61857]: _type = "Task" [ 1005.410996] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.422467] env[61857]: DEBUG oslo_vmware.api [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951470, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.641406] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951467, 'name': Rename_Task, 'duration_secs': 0.161832} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.641763] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1005.642105] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0a5b856-84f3-4e05-a9f1-34e273ac0ed1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.659657] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1005.659657] env[61857]: value = "task-951471" [ 1005.659657] env[61857]: _type = "Task" [ 1005.659657] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.668187] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951471, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.806307] env[61857]: DEBUG oslo_vmware.api [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951468, 'name': PowerOffVM_Task, 'duration_secs': 0.20403} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.806776] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1005.807038] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1005.807330] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-665a614d-3158-47e7-8bdb-098622451f4b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.876466] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1005.876746] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1005.876975] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleting the datastore file [datastore2] f982f0c5-c701-4e5b-b96d-c21a3c0da24f {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.877287] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e487c5c-3094-4470-bdc8-5b3f7bf1cb3a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.884423] env[61857]: DEBUG oslo_vmware.api [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1005.884423] env[61857]: value = "task-951473" [ 1005.884423] env[61857]: _type = "Task" [ 1005.884423] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.895893] env[61857]: DEBUG oslo_vmware.api [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951473, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.922375] env[61857]: DEBUG oslo_vmware.api [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951470, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199232} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.922706] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1005.922918] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1005.923147] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1005.923352] env[61857]: INFO nova.compute.manager [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1005.923632] env[61857]: DEBUG oslo.service.loopingcall [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.923849] env[61857]: DEBUG nova.compute.manager [-] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1005.923946] env[61857]: DEBUG nova.network.neutron [-] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1006.175095] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951471, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.395892] env[61857]: DEBUG oslo_vmware.api [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951473, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.240395} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.396279] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.396522] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1006.396710] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1006.396917] env[61857]: INFO nova.compute.manager [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1006.397276] env[61857]: DEBUG oslo.service.loopingcall [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.397500] env[61857]: DEBUG nova.compute.manager [-] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1006.397593] env[61857]: DEBUG nova.network.neutron [-] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1006.442792] env[61857]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port c646ad3b-3087-4720-bac6-4c9b0108f02e could not be found.", "detail": ""}} {{(pid=61857) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1006.443044] env[61857]: DEBUG nova.network.neutron [-] Unable to show port c646ad3b-3087-4720-bac6-4c9b0108f02e as it no longer exists. {{(pid=61857) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1006.673856] env[61857]: DEBUG oslo_vmware.api [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951471, 'name': PowerOnVM_Task, 'duration_secs': 0.666355} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.674259] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1006.674514] env[61857]: INFO nova.compute.manager [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Took 9.41 seconds to spawn the instance on the hypervisor. [ 1006.674733] env[61857]: DEBUG nova.compute.manager [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.677249] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794dd843-b57c-429a-b418-5e44379741f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.682383] env[61857]: DEBUG nova.network.neutron [-] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.190884] env[61857]: INFO nova.compute.manager [-] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Took 1.27 seconds to deallocate network for instance. [ 1007.207978] env[61857]: INFO nova.compute.manager [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Took 14.29 seconds to build instance. [ 1007.336764] env[61857]: DEBUG nova.network.neutron [-] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.697445] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.697735] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.697972] env[61857]: DEBUG nova.objects.instance [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lazy-loading 'resources' on Instance uuid ce67a15a-8604-4523-a8fa-a34ccf9914da {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.710136] env[61857]: DEBUG oslo_concurrency.lockutils [None req-53c34a43-dc45-4af3-814c-5a125784609a tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.800s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.840545] env[61857]: INFO nova.compute.manager [-] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Took 1.44 seconds to deallocate network for instance. [ 1008.323601] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.323921] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.335768] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede32a97-066a-4bcd-b705-70d931bee3c8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.345047] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa9ab92-e358-4edb-a2bb-0f6e7a639748 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.349677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.381882] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f6e550-fe94-432e-b8cb-8c0620005c75 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.393029] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05fba4e5-ea7d-49e5-9beb-471251bd5145 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.407392] env[61857]: DEBUG nova.compute.provider_tree [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.832498] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.832498] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1008.912068] env[61857]: DEBUG nova.scheduler.client.report [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.417277] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.719s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.419982] env[61857]: DEBUG oslo_concurrency.lockutils [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.070s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.420340] env[61857]: DEBUG nova.objects.instance [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'resources' on Instance uuid f982f0c5-c701-4e5b-b96d-c21a3c0da24f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.440762] env[61857]: INFO nova.scheduler.client.report [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted allocations for instance ce67a15a-8604-4523-a8fa-a34ccf9914da [ 1009.866643] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "refresh_cache-18d9307d-6a20-47b4-b4e9-176f9a8c33cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.866831] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquired lock "refresh_cache-18d9307d-6a20-47b4-b4e9-176f9a8c33cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.866997] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Forcefully refreshing network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1009.948418] env[61857]: DEBUG oslo_concurrency.lockutils [None req-37ee704b-5b46-4ff1-9bca-a681c10c6e70 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "ce67a15a-8604-4523-a8fa-a34ccf9914da" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.207s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.986329] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdb239-3838-2467-9cb9-429242cbd8e6/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1009.987567] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bcf2871-3d40-47fd-8754-eabf60010351 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.999160] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdb239-3838-2467-9cb9-429242cbd8e6/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1009.999160] env[61857]: ERROR oslo_vmware.rw_handles [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdb239-3838-2467-9cb9-429242cbd8e6/disk-0.vmdk due to incomplete transfer. [ 1009.999160] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1e45a0f1-d0b5-4ca1-aba7-42f648378840 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.007606] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cdb239-3838-2467-9cb9-429242cbd8e6/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1010.007682] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Uploaded image 9ed0a868-4498-44da-a97c-65b2330033f4 to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1010.011140] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1010.011140] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-48b58eed-315a-4097-b3c9-51d9d9213a2c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.018915] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 1010.018915] env[61857]: value = "task-951474" [ 1010.018915] env[61857]: _type = "Task" [ 1010.018915] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.020126] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eaeb84a-7464-4d9e-a7da-a3e9e9efe965 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.038068] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624d4b04-ad8d-4665-9c09-ddddccbf3b30 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.040370] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951474, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.074464] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-040db523-9db3-42e2-9ef8-1d9adbc09445 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.082376] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37bf3d9-4f3d-481e-b93b-013b451f2c69 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.097456] env[61857]: DEBUG nova.compute.provider_tree [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.332178] env[61857]: DEBUG nova.compute.manager [req-f0d9aa8e-e739-4527-9657-1555501c05ee req-948932d6-eecf-4edd-92e6-c536a367f874 service nova] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Received event network-vif-deleted-971c8b59-b659-4d32-9fba-17fdfffd9fe2 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.528993] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951474, 'name': Destroy_Task, 'duration_secs': 0.421479} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.529330] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Destroyed the VM [ 1010.529531] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1010.529798] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3ec51b8a-6d5f-4f5f-9ebc-d3221d70ab71 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.536744] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 1010.536744] env[61857]: value = "task-951475" [ 1010.536744] env[61857]: _type = "Task" [ 1010.536744] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.545896] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951475, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.601329] env[61857]: DEBUG nova.scheduler.client.report [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.946382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.946737] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.947054] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.947273] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.947481] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.949883] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "857dbf02-d829-4476-9094-13a30e14c799" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.950176] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.951547] env[61857]: INFO nova.compute.manager [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Terminating instance [ 1010.956874] env[61857]: DEBUG nova.compute.manager [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.957155] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.958082] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60667f71-fe3c-42d8-a5ad-a5e59a07efaf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.968668] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1010.968988] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3bfc8fb-891f-4b1f-9a93-fd576490f6d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.977049] env[61857]: DEBUG oslo_vmware.api [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 1010.977049] env[61857]: value = "task-951476" [ 1010.977049] env[61857]: _type = "Task" [ 1010.977049] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.991754] env[61857]: DEBUG oslo_vmware.api [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951476, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.048789] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951475, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.106461] env[61857]: DEBUG oslo_concurrency.lockutils [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.686s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.128595] env[61857]: INFO nova.scheduler.client.report [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleted allocations for instance f982f0c5-c701-4e5b-b96d-c21a3c0da24f [ 1011.216608] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Updating instance_info_cache with network_info: [{"id": "527bd220-3cfd-4df1-b3c6-014e4bd05cc9", "address": "fa:16:3e:16:4d:0c", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap527bd220-3c", "ovs_interfaceid": "527bd220-3cfd-4df1-b3c6-014e4bd05cc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.310476] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.310788] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.310986] env[61857]: DEBUG nova.compute.manager [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1011.311963] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101e6487-ebfc-4916-8bba-316552c9acc9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.319125] env[61857]: DEBUG nova.compute.manager [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61857) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1011.320146] env[61857]: DEBUG nova.objects.instance [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lazy-loading 'flavor' on Instance uuid 94720bbd-a5db-45ca-9cfc-02c0f127c8bc {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.374447] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "91bb8c20-b7b6-4994-8a0a-b91935397036" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.374764] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "91bb8c20-b7b6-4994-8a0a-b91935397036" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.455154] env[61857]: DEBUG nova.compute.manager [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.487203] env[61857]: DEBUG oslo_vmware.api [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951476, 'name': PowerOffVM_Task, 'duration_secs': 0.346363} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.487512] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1011.487695] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1011.487955] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2153d2de-07ac-4296-a0ca-2bd8721f7fc5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.546164] env[61857]: DEBUG oslo_vmware.api [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951475, 'name': RemoveSnapshot_Task, 'duration_secs': 0.664316} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.547053] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1011.547053] env[61857]: INFO nova.compute.manager [None req-3f66edba-d529-43a9-8c32-d87d3938f4dc tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Took 13.88 seconds to snapshot the instance on the hypervisor. [ 1011.608866] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1011.609109] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1011.609314] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Deleting the datastore file [datastore2] 28bf59ca-4ffe-4005-9a88-da0660ebb48a {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.609594] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7e900c9-68d6-442f-8d7a-b3876d47f691 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.616285] env[61857]: DEBUG oslo_vmware.api [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for the task: (returnval){ [ 1011.616285] env[61857]: value = "task-951478" [ 1011.616285] env[61857]: _type = "Task" [ 1011.616285] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.625894] env[61857]: DEBUG oslo_vmware.api [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951478, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.636960] env[61857]: DEBUG oslo_concurrency.lockutils [None req-52269ef2-821e-4b38-804c-af7ea1ad99df tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.375s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.637849] env[61857]: DEBUG oslo_concurrency.lockutils [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] Acquired lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.638830] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effd38b8-f9f1-4996-a809-25b1d7fc9d4d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.646661] env[61857]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1011.647231] env[61857]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61857) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1011.649282] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-061c3d54-39ea-4b09-b9ce-8729cd10029d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.657981] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dedccb13-764c-448f-ad00-61a239ba0250 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.670082] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217acb-3c7c-c3f0-293a-db11752dadc7/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1011.671346] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f221d4-db67-4fcd-9be6-720b26fb3bba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.678309] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217acb-3c7c-c3f0-293a-db11752dadc7/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1011.678491] env[61857]: ERROR oslo_vmware.rw_handles [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217acb-3c7c-c3f0-293a-db11752dadc7/disk-0.vmdk due to incomplete transfer. [ 1011.678944] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-94fd3867-45f0-445c-b08e-4bd81ed98165 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.689866] env[61857]: ERROR root [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-214163' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-214163' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-214163' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-214163'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-214163' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-214163' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-214163'}\n"]: nova.exception.InstanceNotFound: Instance f982f0c5-c701-4e5b-b96d-c21a3c0da24f could not be found. [ 1011.690093] env[61857]: DEBUG oslo_concurrency.lockutils [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] Releasing lock "f982f0c5-c701-4e5b-b96d-c21a3c0da24f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.690353] env[61857]: DEBUG nova.compute.manager [req-82bb113d-6434-40b9-82fb-4df5453f51f1 req-bc2dcf54-7a8b-46ca-ade4-c9f16e62b5d4 service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Detach interface failed, port_id=d306c021-073e-45f7-9916-f3a6bc7a7dff, reason: Instance f982f0c5-c701-4e5b-b96d-c21a3c0da24f could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1011.691540] env[61857]: DEBUG oslo_vmware.rw_handles [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52217acb-3c7c-c3f0-293a-db11752dadc7/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1011.691799] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Uploaded image 6f2ff642-ffbc-477c-83d9-8ce6daf6b818 to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1011.693375] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1011.693652] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0a9dfce5-b24b-482e-accf-f0615b334832 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.699273] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1011.699273] env[61857]: value = "task-951479" [ 1011.699273] env[61857]: _type = "Task" [ 1011.699273] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.706995] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951479, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.720038] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Releasing lock "refresh_cache-18d9307d-6a20-47b4-b4e9-176f9a8c33cc" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.720038] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Updated the network info_cache for instance {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1011.720038] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1011.720279] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1011.720467] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1011.720631] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1011.720809] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1011.720999] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1011.721558] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1011.721558] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1011.826238] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1011.826518] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc94e400-ed13-414b-bfa9-54cc6e942632 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.836271] env[61857]: DEBUG oslo_vmware.api [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1011.836271] env[61857]: value = "task-951480" [ 1011.836271] env[61857]: _type = "Task" [ 1011.836271] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.844150] env[61857]: DEBUG oslo_vmware.api [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.869931] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.870532] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.877550] env[61857]: DEBUG nova.compute.manager [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.983265] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.983548] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.985532] env[61857]: INFO nova.compute.claims [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.130034] env[61857]: DEBUG oslo_vmware.api [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Task: {'id': task-951478, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179441} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.130775] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.130974] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1012.131170] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1012.131400] env[61857]: INFO nova.compute.manager [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1012.131655] env[61857]: DEBUG oslo.service.loopingcall [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.131857] env[61857]: DEBUG nova.compute.manager [-] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1012.131950] env[61857]: DEBUG nova.network.neutron [-] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1012.209670] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951479, 'name': Destroy_Task, 'duration_secs': 0.352601} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.209951] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Destroyed the VM [ 1012.210377] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1012.210603] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b1790065-e124-4c3c-9c26-032ba3a6681b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.217191] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1012.217191] env[61857]: value = "task-951481" [ 1012.217191] env[61857]: _type = "Task" [ 1012.217191] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.225769] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.226127] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.345809] env[61857]: DEBUG oslo_vmware.api [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951480, 'name': PowerOffVM_Task, 'duration_secs': 0.214906} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.347034] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1012.347034] env[61857]: DEBUG nova.compute.manager [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1012.348021] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b632988d-0e73-4d1e-8000-61b99835e10d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.375646] env[61857]: DEBUG nova.compute.manager [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1012.415948] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.480403] env[61857]: DEBUG nova.compute.manager [req-34f7dab8-4f4d-4565-9f38-24565b2b0928 req-538f268b-ad4c-4b51-aeb0-be395f507aae service nova] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Received event network-vif-deleted-9d3974fc-a398-4220-9f8c-d78aa99796bc {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.726687] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951481, 'name': RemoveSnapshot_Task, 'duration_secs': 0.351407} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.727041] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1012.727328] env[61857]: DEBUG nova.compute.manager [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1012.728131] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bc3c1b-ccf8-43a0-bd18-8152c50a590b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.864522] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7438a501-86bd-4b62-b4c2-5f203e177e99 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.554s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.897069] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.897069] env[61857]: DEBUG nova.network.neutron [-] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.112838] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87c36d2-7ceb-4a58-b56f-57b16e6844a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.120419] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18eab603-d008-4a7c-9d72-3202a65a9cec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.151111] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbd6304-7f9d-4ccf-a282-b5f29761b135 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.158673] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8627f9-e2b7-4a4a-b04a-8f7f96b5dc4f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.171870] env[61857]: DEBUG nova.compute.provider_tree [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.241210] env[61857]: INFO nova.compute.manager [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Shelve offloading [ 1013.243333] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1013.243515] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-22b2e460-5118-4edc-9fd4-ed6f532b9ce5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.250213] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1013.250213] env[61857]: value = "task-951482" [ 1013.250213] env[61857]: _type = "Task" [ 1013.250213] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.258041] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951482, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.399443] env[61857]: DEBUG nova.compute.manager [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.399830] env[61857]: INFO nova.compute.manager [-] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Took 1.27 seconds to deallocate network for instance. [ 1013.400727] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d8a3f2-569a-45c6-8527-cf2ab162f940 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.465806] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "14844876-96ab-481d-980d-4133e89d7d4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.466094] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14844876-96ab-481d-980d-4133e89d7d4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.675443] env[61857]: DEBUG nova.scheduler.client.report [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.760606] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] VM already powered off {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1013.760934] env[61857]: DEBUG nova.compute.manager [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.761736] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6b114a-c422-4782-8bd8-c68b8d2808f8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.767497] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.767665] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.767834] env[61857]: DEBUG nova.network.neutron [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1013.814586] env[61857]: DEBUG oslo_concurrency.lockutils [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.814848] env[61857]: DEBUG oslo_concurrency.lockutils [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.815071] env[61857]: DEBUG oslo_concurrency.lockutils [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.815269] env[61857]: DEBUG oslo_concurrency.lockutils [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.815446] env[61857]: DEBUG oslo_concurrency.lockutils [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.817450] env[61857]: INFO nova.compute.manager [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Terminating instance [ 1013.819316] env[61857]: DEBUG nova.compute.manager [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1013.819523] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1013.820351] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02896866-dbc3-4e5b-9112-7fc23df84067 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.827201] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1013.827426] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b442c6c8-819f-44a2-b4a1-f2d847a2457e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.913396] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.914372] env[61857]: INFO nova.compute.manager [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] instance snapshotting [ 1013.917161] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1013.917380] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1013.917563] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleting the datastore file [datastore1] 94720bbd-a5db-45ca-9cfc-02c0f127c8bc {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.918110] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d170dfcb-fb92-4372-8d5f-53d2920ad267 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.920548] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2539518-6632-4959-97a8-ebc091b94f32 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.939480] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585fe72a-3f12-4313-979f-2aa199527b69 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.942119] env[61857]: DEBUG oslo_vmware.api [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1013.942119] env[61857]: value = "task-951484" [ 1013.942119] env[61857]: _type = "Task" [ 1013.942119] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.953185] env[61857]: DEBUG oslo_vmware.api [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951484, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.967669] env[61857]: DEBUG nova.compute.manager [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1014.180178] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.196s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.180748] env[61857]: DEBUG nova.compute.manager [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1014.183550] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.958s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.183734] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.183890] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1014.184220] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.768s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.185614] env[61857]: INFO nova.compute.claims [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.188630] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08008ce1-3f32-4f10-a7bf-6d94acb2ebf5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.197290] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea87bfb-af11-40f2-a588-fc8226db3446 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.211119] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282b7dfb-9879-4616-aaba-e7ba6981a775 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.217763] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e13d61-8f21-4f5d-a963-cca5136e496f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.249546] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180033MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1014.249691] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.453414] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1014.453744] env[61857]: DEBUG oslo_vmware.api [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951484, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148429} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.453969] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8cbd6542-2291-4d48-8529-b30dd68e2a78 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.455906] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.456122] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1014.456311] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1014.456494] env[61857]: INFO nova.compute.manager [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Took 0.64 seconds to destroy the instance on the hypervisor. [ 1014.456741] env[61857]: DEBUG oslo.service.loopingcall [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.459013] env[61857]: DEBUG nova.compute.manager [-] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.459125] env[61857]: DEBUG nova.network.neutron [-] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1014.465968] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 1014.465968] env[61857]: value = "task-951485" [ 1014.465968] env[61857]: _type = "Task" [ 1014.465968] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.476478] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951485, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.503482] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.512538] env[61857]: DEBUG nova.compute.manager [req-04a5361c-1c1a-4077-a859-b8db3f636d82 req-f9b5e06a-5f11-4e62-a4e0-653464c78b3d service nova] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Received event network-vif-deleted-45483cb0-6223-4cc6-8f9c-b87c3d5fa7d6 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.535012] env[61857]: DEBUG nova.network.neutron [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating instance_info_cache with network_info: [{"id": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "address": "fa:16:3e:52:2f:3f", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507983f7-c4", "ovs_interfaceid": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.690289] env[61857]: DEBUG nova.compute.utils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.691729] env[61857]: DEBUG nova.compute.manager [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1014.692136] env[61857]: DEBUG nova.network.neutron [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1014.772932] env[61857]: DEBUG nova.policy [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4c9ee17921646978e8b21ad5508813c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2a7d2f228f88485f863b3c421b25a37c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1014.976126] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951485, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.037814] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.194873] env[61857]: DEBUG nova.compute.manager [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1015.309555] env[61857]: DEBUG nova.network.neutron [-] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.318564] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d844960-b7ca-4c19-ad3c-2ba33c1bfe82 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.326691] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630e45f7-2732-4993-bc6d-2d0c97d9e63a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.360060] env[61857]: DEBUG nova.network.neutron [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Successfully created port: f3163567-eee7-4192-9332-fe1748882d93 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1015.362362] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75af8f3e-e90f-448f-8aba-103fe4f5afc0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.370051] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698d5316-3d07-4ce8-ab5b-09114e81945f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.384281] env[61857]: DEBUG nova.compute.provider_tree [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.476489] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951485, 'name': CreateSnapshot_Task, 'duration_secs': 0.532468} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.476777] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1015.477532] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb19a0e-e214-4b39-bb1e-8713f83e479e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.812579] env[61857]: INFO nova.compute.manager [-] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Took 1.35 seconds to deallocate network for instance. [ 1015.861886] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1015.862813] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95663a05-61e0-4c00-b478-214a5c6ee39f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.870750] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1015.871123] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4766440-5417-4e55-9291-6c3fd51d3d33 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.888221] env[61857]: DEBUG nova.scheduler.client.report [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.941021] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1015.941309] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1015.941536] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleting the datastore file [datastore2] a42e8c77-3cc3-4323-b1ee-2a95192910d5 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.941812] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3f5d5aeb-d567-4101-a426-a2733536f08a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.948200] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1015.948200] env[61857]: value = "task-951487" [ 1015.948200] env[61857]: _type = "Task" [ 1015.948200] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.956996] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951487, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.994495] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1015.994801] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b2c4c65e-11b4-40b0-b95c-db9c3a59fc9e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.002466] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 1016.002466] env[61857]: value = "task-951488" [ 1016.002466] env[61857]: _type = "Task" [ 1016.002466] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.010449] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951488, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.207579] env[61857]: DEBUG nova.compute.manager [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1016.230882] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.231128] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.231308] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.231496] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.231657] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.231840] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.232074] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.232247] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.232422] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.232590] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.232766] env[61857]: DEBUG nova.virt.hardware [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.233646] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910c19c1-e991-40fc-b3c9-1f8b5adc4d09 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.242787] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f773693-f6cc-4bee-ac7b-d32d4aec5e56 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.323919] env[61857]: DEBUG oslo_concurrency.lockutils [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.392100] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.392669] env[61857]: DEBUG nova.compute.manager [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1016.395291] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.499s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.396717] env[61857]: INFO nova.compute.claims [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.458254] env[61857]: DEBUG oslo_vmware.api [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951487, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138825} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.458601] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.458812] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1016.459017] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1016.479465] env[61857]: INFO nova.scheduler.client.report [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleted allocations for instance a42e8c77-3cc3-4323-b1ee-2a95192910d5 [ 1016.512238] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951488, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.541401] env[61857]: DEBUG nova.compute.manager [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Received event network-vif-deleted-33d4e605-4d10-46de-8fe5-af0c42295576 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.541484] env[61857]: DEBUG nova.compute.manager [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received event network-vif-unplugged-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.541687] env[61857]: DEBUG oslo_concurrency.lockutils [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] Acquiring lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.541895] env[61857]: DEBUG oslo_concurrency.lockutils [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.542080] env[61857]: DEBUG oslo_concurrency.lockutils [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.542254] env[61857]: DEBUG nova.compute.manager [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] No waiting events found dispatching network-vif-unplugged-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1016.542432] env[61857]: WARNING nova.compute.manager [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received unexpected event network-vif-unplugged-507983f7-c4c5-4605-887e-c3a1936e8f3a for instance with vm_state shelved_offloaded and task_state None. [ 1016.542601] env[61857]: DEBUG nova.compute.manager [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received event network-changed-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.542760] env[61857]: DEBUG nova.compute.manager [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Refreshing instance network info cache due to event network-changed-507983f7-c4c5-4605-887e-c3a1936e8f3a. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1016.542955] env[61857]: DEBUG oslo_concurrency.lockutils [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] Acquiring lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.543102] env[61857]: DEBUG oslo_concurrency.lockutils [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] Acquired lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.543268] env[61857]: DEBUG nova.network.neutron [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Refreshing network info cache for port 507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1016.901100] env[61857]: DEBUG nova.compute.utils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1016.904650] env[61857]: DEBUG nova.compute.manager [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1016.904826] env[61857]: DEBUG nova.network.neutron [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1016.984484] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.012293] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951488, 'name': CloneVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.018181] env[61857]: DEBUG nova.policy [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2ef7e845007475a8a19cf3f520c3f51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1acf476aabc4166bc8505a3442367c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1017.237721] env[61857]: DEBUG nova.network.neutron [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Successfully updated port: f3163567-eee7-4192-9332-fe1748882d93 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1017.306395] env[61857]: DEBUG nova.network.neutron [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updated VIF entry in instance network info cache for port 507983f7-c4c5-4605-887e-c3a1936e8f3a. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.306745] env[61857]: DEBUG nova.network.neutron [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating instance_info_cache with network_info: [{"id": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "address": "fa:16:3e:52:2f:3f", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap507983f7-c4", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.378636] env[61857]: DEBUG nova.network.neutron [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Successfully created port: 3a0cf2fa-0204-4f9e-8b78-2576d4e8524f {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1017.405172] env[61857]: DEBUG nova.compute.manager [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1017.515923] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951488, 'name': CloneVM_Task, 'duration_secs': 1.07105} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.516549] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Created linked-clone VM from snapshot [ 1017.520358] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbe9206-bb7b-45b3-9332-dd31ad08f1ed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.529321] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Uploading image 2382b3d3-d1ff-4aa0-90ba-7e846851d4d5 {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1017.557947] env[61857]: DEBUG oslo_vmware.rw_handles [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1017.557947] env[61857]: value = "vm-214199" [ 1017.557947] env[61857]: _type = "VirtualMachine" [ 1017.557947] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1017.558256] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a5d8c0bf-dc7c-4b2e-b7f7-0e62b50755ad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.564527] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6115f9d-7695-4043-a87e-a4a5c755059f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.568104] env[61857]: DEBUG oslo_vmware.rw_handles [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lease: (returnval){ [ 1017.568104] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ead013-35bd-ba26-14ef-596a3dca5bfc" [ 1017.568104] env[61857]: _type = "HttpNfcLease" [ 1017.568104] env[61857]: } obtained for exporting VM: (result){ [ 1017.568104] env[61857]: value = "vm-214199" [ 1017.568104] env[61857]: _type = "VirtualMachine" [ 1017.568104] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1017.568377] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the lease: (returnval){ [ 1017.568377] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ead013-35bd-ba26-14ef-596a3dca5bfc" [ 1017.568377] env[61857]: _type = "HttpNfcLease" [ 1017.568377] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1017.574605] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9fc5ad-85f4-4773-98ed-a798072ecbdf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.579329] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1017.579329] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ead013-35bd-ba26-14ef-596a3dca5bfc" [ 1017.579329] env[61857]: _type = "HttpNfcLease" [ 1017.579329] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1017.607746] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52faf9d-9cc9-4e1d-acf2-fe6410843436 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.615398] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12131560-68dd-4970-be1a-decd25dee5a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.628774] env[61857]: DEBUG nova.compute.provider_tree [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.741890] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.742076] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.742229] env[61857]: DEBUG nova.network.neutron [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1017.810810] env[61857]: DEBUG oslo_concurrency.lockutils [req-e8b557ed-be4b-498b-9b72-24f584c910b3 req-58cb17d0-cc01-4be4-831a-3b8490e166b2 service nova] Releasing lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.077520] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1018.077520] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ead013-35bd-ba26-14ef-596a3dca5bfc" [ 1018.077520] env[61857]: _type = "HttpNfcLease" [ 1018.077520] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1018.077907] env[61857]: DEBUG oslo_vmware.rw_handles [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1018.077907] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ead013-35bd-ba26-14ef-596a3dca5bfc" [ 1018.077907] env[61857]: _type = "HttpNfcLease" [ 1018.077907] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1018.078630] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f74a67d-7e61-4930-8ce0-be88db13fa09 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.086112] env[61857]: DEBUG oslo_vmware.rw_handles [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961971-94de-f848-cfbd-ec2a268ca36e/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1018.086299] env[61857]: DEBUG oslo_vmware.rw_handles [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961971-94de-f848-cfbd-ec2a268ca36e/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1018.142856] env[61857]: DEBUG nova.scheduler.client.report [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.179328] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f5ef72e9-70d0-4891-a5c2-ef1b33c5e0f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.275954] env[61857]: DEBUG nova.network.neutron [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1018.417384] env[61857]: DEBUG nova.compute.manager [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1018.432299] env[61857]: DEBUG nova.network.neutron [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance_info_cache with network_info: [{"id": "f3163567-eee7-4192-9332-fe1748882d93", "address": "fa:16:3e:80:c4:35", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3163567-ee", "ovs_interfaceid": "f3163567-eee7-4192-9332-fe1748882d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.442257] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.442532] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.442702] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.442896] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.443070] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.443228] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.443440] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.443607] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.443781] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.443951] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.444145] env[61857]: DEBUG nova.virt.hardware [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.445025] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acab2070-a9c0-4b3e-925b-65c97e19d933 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.454116] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820d7d0c-c9ef-4109-89d9-3482b178a2d0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.567485] env[61857]: DEBUG nova.compute.manager [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Received event network-vif-plugged-f3163567-eee7-4192-9332-fe1748882d93 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.567870] env[61857]: DEBUG oslo_concurrency.lockutils [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] Acquiring lock "857dbf02-d829-4476-9094-13a30e14c799-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.568203] env[61857]: DEBUG oslo_concurrency.lockutils [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] Lock "857dbf02-d829-4476-9094-13a30e14c799-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.568494] env[61857]: DEBUG oslo_concurrency.lockutils [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] Lock "857dbf02-d829-4476-9094-13a30e14c799-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.568778] env[61857]: DEBUG nova.compute.manager [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] No waiting events found dispatching network-vif-plugged-f3163567-eee7-4192-9332-fe1748882d93 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.568961] env[61857]: WARNING nova.compute.manager [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Received unexpected event network-vif-plugged-f3163567-eee7-4192-9332-fe1748882d93 for instance with vm_state building and task_state spawning. [ 1018.569149] env[61857]: DEBUG nova.compute.manager [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Received event network-changed-f3163567-eee7-4192-9332-fe1748882d93 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.569327] env[61857]: DEBUG nova.compute.manager [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Refreshing instance network info cache due to event network-changed-f3163567-eee7-4192-9332-fe1748882d93. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1018.569571] env[61857]: DEBUG oslo_concurrency.lockutils [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] Acquiring lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.653255] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.257s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.653969] env[61857]: DEBUG nova.compute.manager [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1018.656693] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.743s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.657118] env[61857]: DEBUG nova.objects.instance [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lazy-loading 'resources' on Instance uuid 28bf59ca-4ffe-4005-9a88-da0660ebb48a {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.658350] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.940024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.940024] env[61857]: DEBUG nova.compute.manager [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Instance network_info: |[{"id": "f3163567-eee7-4192-9332-fe1748882d93", "address": "fa:16:3e:80:c4:35", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3163567-ee", "ovs_interfaceid": "f3163567-eee7-4192-9332-fe1748882d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1018.940024] env[61857]: DEBUG oslo_concurrency.lockutils [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] Acquired lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.940024] env[61857]: DEBUG nova.network.neutron [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Refreshing network info cache for port f3163567-eee7-4192-9332-fe1748882d93 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1018.940024] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:c4:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '170f3b82-5915-4e36-bce9-4664ebb6be5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f3163567-eee7-4192-9332-fe1748882d93', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.957361] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating folder: Project (2a7d2f228f88485f863b3c421b25a37c). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1018.961946] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8a605288-acf0-464b-9eb6-53e4adaef692 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.975578] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Created folder: Project (2a7d2f228f88485f863b3c421b25a37c) in parent group-v214027. [ 1018.975578] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating folder: Instances. Parent ref: group-v214200. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1018.975578] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-efd79ee0-91ba-4374-9e6a-8bd24801be61 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.985579] env[61857]: DEBUG nova.network.neutron [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Successfully updated port: 3a0cf2fa-0204-4f9e-8b78-2576d4e8524f {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.987731] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Created folder: Instances in parent group-v214200. [ 1018.988123] env[61857]: DEBUG oslo.service.loopingcall [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.988311] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1018.988559] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ba5395e-78cb-4b1d-871e-b93fcdb1e82f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.012788] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.012788] env[61857]: value = "task-951492" [ 1019.012788] env[61857]: _type = "Task" [ 1019.012788] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.021047] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951492, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.163540] env[61857]: DEBUG nova.compute.utils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1019.165603] env[61857]: DEBUG nova.compute.manager [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1019.165603] env[61857]: DEBUG nova.network.neutron [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1019.197092] env[61857]: DEBUG nova.network.neutron [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updated VIF entry in instance network info cache for port f3163567-eee7-4192-9332-fe1748882d93. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1019.197482] env[61857]: DEBUG nova.network.neutron [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance_info_cache with network_info: [{"id": "f3163567-eee7-4192-9332-fe1748882d93", "address": "fa:16:3e:80:c4:35", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3163567-ee", "ovs_interfaceid": "f3163567-eee7-4192-9332-fe1748882d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.240695] env[61857]: DEBUG nova.policy [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f719cbf1755d4c34bef92dab6679cb56', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ef111c8b3ff4f2383b4e81db026792b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1019.315876] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7502054-3db4-40b3-ae9b-058299dd0200 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.325077] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b735119-2888-4405-8cfa-54c1c53f7668 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.359151] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103686ac-7fd8-4ef8-b522-2c0b36936028 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.367866] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9af1cfa-e103-427e-ae0e-aeff3e5f7d06 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.381800] env[61857]: DEBUG nova.compute.provider_tree [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.489183] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "refresh_cache-91bb8c20-b7b6-4994-8a0a-b91935397036" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.489297] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "refresh_cache-91bb8c20-b7b6-4994-8a0a-b91935397036" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.489485] env[61857]: DEBUG nova.network.neutron [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1019.522350] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951492, 'name': CreateVM_Task, 'duration_secs': 0.326349} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.522586] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1019.523289] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.523528] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.524007] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1019.524378] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e67c4d4a-5911-4f7f-a91f-46691a323682 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.529126] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1019.529126] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52c6665f-af5d-f9f1-39c5-9bc1d36ab431" [ 1019.529126] env[61857]: _type = "Task" [ 1019.529126] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.536750] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c6665f-af5d-f9f1-39c5-9bc1d36ab431, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.590582] env[61857]: DEBUG nova.network.neutron [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Successfully created port: 83ff6e16-8b1b-4391-9ef1-57600502688c {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1019.668790] env[61857]: DEBUG nova.compute.manager [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1019.703686] env[61857]: DEBUG oslo_concurrency.lockutils [req-f9b32ede-264e-46f3-8b91-18ff4d8f411c req-f4f94618-0b41-4d93-8ca6-f467605942c8 service nova] Releasing lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.885178] env[61857]: DEBUG nova.scheduler.client.report [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.019873] env[61857]: DEBUG nova.network.neutron [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1020.043769] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52c6665f-af5d-f9f1-39c5-9bc1d36ab431, 'name': SearchDatastore_Task, 'duration_secs': 0.011222} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.044161] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.044441] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1020.044696] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.044863] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.045064] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.045355] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d965b6d-bcfe-4a96-803e-ae46abe52171 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.054673] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.054992] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1020.055929] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb573acf-0091-4dac-b0c7-05797e197916 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.062683] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1020.062683] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52413a51-f1fd-90be-ee7b-769367a3729d" [ 1020.062683] env[61857]: _type = "Task" [ 1020.062683] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.074554] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52413a51-f1fd-90be-ee7b-769367a3729d, 'name': SearchDatastore_Task, 'duration_secs': 0.008251} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.075451] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09274eaa-ca99-4e1a-93f5-5407de559f8e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.080952] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1020.080952] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e0b19b-e8fd-97cb-4db8-22c6f91536e1" [ 1020.080952] env[61857]: _type = "Task" [ 1020.080952] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.088821] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e0b19b-e8fd-97cb-4db8-22c6f91536e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.178815] env[61857]: DEBUG nova.network.neutron [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Updating instance_info_cache with network_info: [{"id": "3a0cf2fa-0204-4f9e-8b78-2576d4e8524f", "address": "fa:16:3e:31:83:e1", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a0cf2fa-02", "ovs_interfaceid": "3a0cf2fa-0204-4f9e-8b78-2576d4e8524f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.390476] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.734s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.393043] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.143s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.414348] env[61857]: INFO nova.scheduler.client.report [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Deleted allocations for instance 28bf59ca-4ffe-4005-9a88-da0660ebb48a [ 1020.591928] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e0b19b-e8fd-97cb-4db8-22c6f91536e1, 'name': SearchDatastore_Task, 'duration_secs': 0.008772} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.593034] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.593034] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 857dbf02-d829-4476-9094-13a30e14c799/857dbf02-d829-4476-9094-13a30e14c799.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1020.594168] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c60325a-58fa-4ce9-9e24-d0f2b6cf1ee5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.597529] env[61857]: DEBUG nova.compute.manager [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Received event network-vif-plugged-3a0cf2fa-0204-4f9e-8b78-2576d4e8524f {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.597751] env[61857]: DEBUG oslo_concurrency.lockutils [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] Acquiring lock "91bb8c20-b7b6-4994-8a0a-b91935397036-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.597971] env[61857]: DEBUG oslo_concurrency.lockutils [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] Lock "91bb8c20-b7b6-4994-8a0a-b91935397036-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.598208] env[61857]: DEBUG oslo_concurrency.lockutils [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] Lock "91bb8c20-b7b6-4994-8a0a-b91935397036-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.598408] env[61857]: DEBUG nova.compute.manager [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] No waiting events found dispatching network-vif-plugged-3a0cf2fa-0204-4f9e-8b78-2576d4e8524f {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.598601] env[61857]: WARNING nova.compute.manager [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Received unexpected event network-vif-plugged-3a0cf2fa-0204-4f9e-8b78-2576d4e8524f for instance with vm_state building and task_state spawning. [ 1020.598817] env[61857]: DEBUG nova.compute.manager [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Received event network-changed-3a0cf2fa-0204-4f9e-8b78-2576d4e8524f {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.598990] env[61857]: DEBUG nova.compute.manager [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Refreshing instance network info cache due to event network-changed-3a0cf2fa-0204-4f9e-8b78-2576d4e8524f. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1020.599197] env[61857]: DEBUG oslo_concurrency.lockutils [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] Acquiring lock "refresh_cache-91bb8c20-b7b6-4994-8a0a-b91935397036" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.606060] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1020.606060] env[61857]: value = "task-951493" [ 1020.606060] env[61857]: _type = "Task" [ 1020.606060] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.614566] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951493, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.679609] env[61857]: DEBUG nova.compute.manager [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1020.682191] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "refresh_cache-91bb8c20-b7b6-4994-8a0a-b91935397036" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.682484] env[61857]: DEBUG nova.compute.manager [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Instance network_info: |[{"id": "3a0cf2fa-0204-4f9e-8b78-2576d4e8524f", "address": "fa:16:3e:31:83:e1", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a0cf2fa-02", "ovs_interfaceid": "3a0cf2fa-0204-4f9e-8b78-2576d4e8524f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1020.682786] env[61857]: DEBUG oslo_concurrency.lockutils [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] Acquired lock "refresh_cache-91bb8c20-b7b6-4994-8a0a-b91935397036" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.682988] env[61857]: DEBUG nova.network.neutron [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Refreshing network info cache for port 3a0cf2fa-0204-4f9e-8b78-2576d4e8524f {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1020.684267] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:83:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a0cf2fa-0204-4f9e-8b78-2576d4e8524f', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.692317] env[61857]: DEBUG oslo.service.loopingcall [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.695530] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1020.696079] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a3d4ba49-4ba3-4e23-9dbf-47ae95a1b265 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.719984] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.720289] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.720451] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.720650] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.720878] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.721091] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.721284] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.721461] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.721639] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.721795] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.721973] env[61857]: DEBUG nova.virt.hardware [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.723044] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c7cc23-544e-476b-a20b-c59f10cc6c02 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.726770] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.726770] env[61857]: value = "task-951494" [ 1020.726770] env[61857]: _type = "Task" [ 1020.726770] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.737470] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42dedfb4-1aca-4372-bbb5-50502833f320 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.741150] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951494, 'name': CreateVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.928412] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d91ddca1-daf5-474c-890d-17f23484adbc tempest-ServerRescueNegativeTestJSON-205112447 tempest-ServerRescueNegativeTestJSON-205112447-project-member] Lock "28bf59ca-4ffe-4005-9a88-da0660ebb48a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.981s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.971663] env[61857]: DEBUG nova.network.neutron [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Updated VIF entry in instance network info cache for port 3a0cf2fa-0204-4f9e-8b78-2576d4e8524f. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1020.972077] env[61857]: DEBUG nova.network.neutron [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Updating instance_info_cache with network_info: [{"id": "3a0cf2fa-0204-4f9e-8b78-2576d4e8524f", "address": "fa:16:3e:31:83:e1", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a0cf2fa-02", "ovs_interfaceid": "3a0cf2fa-0204-4f9e-8b78-2576d4e8524f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.119463] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951493, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47899} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.119756] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 857dbf02-d829-4476-9094-13a30e14c799/857dbf02-d829-4476-9094-13a30e14c799.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1021.119986] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.120298] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-99fb9673-6c91-4d40-a8a7-07327903fe59 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.126808] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1021.126808] env[61857]: value = "task-951495" [ 1021.126808] env[61857]: _type = "Task" [ 1021.126808] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.134574] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951495, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.237285] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951494, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.248483] env[61857]: DEBUG nova.network.neutron [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Successfully updated port: 83ff6e16-8b1b-4391-9ef1-57600502688c {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1021.426051] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 18d9307d-6a20-47b4-b4e9-176f9a8c33cc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.426240] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.426393] env[61857]: WARNING nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 94720bbd-a5db-45ca-9cfc-02c0f127c8bc is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1021.426520] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 857dbf02-d829-4476-9094-13a30e14c799 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.426639] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 91bb8c20-b7b6-4994-8a0a-b91935397036 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.426755] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 04438fc0-b5a1-4e29-84d5-4d84d332c8e4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.477486] env[61857]: DEBUG oslo_concurrency.lockutils [req-3fbb2454-0499-4276-933f-0bc522fa0dc1 req-aba2fb3b-54a6-4aad-895e-79b3fe97422a service nova] Releasing lock "refresh_cache-91bb8c20-b7b6-4994-8a0a-b91935397036" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.637041] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951495, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062878} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.637371] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.638162] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de4fbff-b03e-4532-8c4a-9804a5bec3bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.659903] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 857dbf02-d829-4476-9094-13a30e14c799/857dbf02-d829-4476-9094-13a30e14c799.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.660211] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8aba281-258d-48e0-a98e-4b984382ec25 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.682076] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1021.682076] env[61857]: value = "task-951496" [ 1021.682076] env[61857]: _type = "Task" [ 1021.682076] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.689770] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951496, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.737797] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951494, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.751545] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.751710] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.751831] env[61857]: DEBUG nova.network.neutron [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.930236] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 14844876-96ab-481d-980d-4133e89d7d4d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1021.930472] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (2d5860d0-8574-4e55-9ced-20e66f0314c2): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 1021.930692] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1021.930838] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1022.055354] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82e2d96-8cff-48f3-843c-566145a7453a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.063633] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce70aa2-be5e-4dac-91dd-e813739277b4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.094397] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44cc6672-0abd-4c52-b728-94eed7b5438b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.102197] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65578b6-1dbf-4214-a4f6-99a04ccd4f5a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.115715] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.193046] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951496, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.238067] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951494, 'name': CreateVM_Task, 'duration_secs': 1.409227} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.238445] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1022.238984] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.239181] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.239518] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1022.239810] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-850d757f-9cea-40bf-8772-e8ba3be7af77 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.245253] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1022.245253] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52600ac8-bd4f-daf7-2162-740328bfc829" [ 1022.245253] env[61857]: _type = "Task" [ 1022.245253] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.257087] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52600ac8-bd4f-daf7-2162-740328bfc829, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.321132] env[61857]: DEBUG nova.network.neutron [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1022.543699] env[61857]: DEBUG nova.network.neutron [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance_info_cache with network_info: [{"id": "83ff6e16-8b1b-4391-9ef1-57600502688c", "address": "fa:16:3e:46:c4:d7", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83ff6e16-8b", "ovs_interfaceid": "83ff6e16-8b1b-4391-9ef1-57600502688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.621772] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.627470] env[61857]: DEBUG nova.compute.manager [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Received event network-vif-plugged-83ff6e16-8b1b-4391-9ef1-57600502688c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.630892] env[61857]: DEBUG oslo_concurrency.lockutils [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] Acquiring lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.630892] env[61857]: DEBUG oslo_concurrency.lockutils [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] Lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.630892] env[61857]: DEBUG oslo_concurrency.lockutils [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] Lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.630892] env[61857]: DEBUG nova.compute.manager [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] No waiting events found dispatching network-vif-plugged-83ff6e16-8b1b-4391-9ef1-57600502688c {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.630892] env[61857]: WARNING nova.compute.manager [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Received unexpected event network-vif-plugged-83ff6e16-8b1b-4391-9ef1-57600502688c for instance with vm_state building and task_state spawning. [ 1022.630892] env[61857]: DEBUG nova.compute.manager [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Received event network-changed-83ff6e16-8b1b-4391-9ef1-57600502688c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.630892] env[61857]: DEBUG nova.compute.manager [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Refreshing instance network info cache due to event network-changed-83ff6e16-8b1b-4391-9ef1-57600502688c. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1022.630892] env[61857]: DEBUG oslo_concurrency.lockutils [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] Acquiring lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.692615] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951496, 'name': ReconfigVM_Task, 'duration_secs': 0.681134} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.693322] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 857dbf02-d829-4476-9094-13a30e14c799/857dbf02-d829-4476-9094-13a30e14c799.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.694478] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1c48929-cbf3-444f-9c71-3cd0c5818a15 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.701311] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1022.701311] env[61857]: value = "task-951497" [ 1022.701311] env[61857]: _type = "Task" [ 1022.701311] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.709354] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951497, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.756819] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52600ac8-bd4f-daf7-2162-740328bfc829, 'name': SearchDatastore_Task, 'duration_secs': 0.009701} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.757353] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.757631] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1022.757827] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.758047] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.758348] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1022.758981] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a31ca4c8-d17c-4379-bdee-86bb97cf366d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.768639] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1022.768833] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1022.769578] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a873118e-99d1-4b50-8b07-79a4b7a277b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.775315] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1022.775315] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52181f76-5ca2-3d87-2625-f1d855421c61" [ 1022.775315] env[61857]: _type = "Task" [ 1022.775315] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.783250] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52181f76-5ca2-3d87-2625-f1d855421c61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.046440] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.046880] env[61857]: DEBUG nova.compute.manager [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Instance network_info: |[{"id": "83ff6e16-8b1b-4391-9ef1-57600502688c", "address": "fa:16:3e:46:c4:d7", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83ff6e16-8b", "ovs_interfaceid": "83ff6e16-8b1b-4391-9ef1-57600502688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1023.047301] env[61857]: DEBUG oslo_concurrency.lockutils [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] Acquired lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.047569] env[61857]: DEBUG nova.network.neutron [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Refreshing network info cache for port 83ff6e16-8b1b-4391-9ef1-57600502688c {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1023.050063] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:c4:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '83ff6e16-8b1b-4391-9ef1-57600502688c', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1023.059386] env[61857]: DEBUG oslo.service.loopingcall [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.062581] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1023.063131] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f267cb3-babc-4b3b-a974-29355346ca98 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.084188] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1023.084188] env[61857]: value = "task-951498" [ 1023.084188] env[61857]: _type = "Task" [ 1023.084188] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.093232] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951498, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.126897] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1023.127198] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.734s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.127915] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.624s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.129245] env[61857]: INFO nova.compute.claims [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.214205] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951497, 'name': Rename_Task, 'duration_secs': 0.151664} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.215049] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1023.215049] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77243c86-6c33-4dfa-ac82-283eb73b15a5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.223293] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1023.223293] env[61857]: value = "task-951499" [ 1023.223293] env[61857]: _type = "Task" [ 1023.223293] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.239892] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951499, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.286629] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52181f76-5ca2-3d87-2625-f1d855421c61, 'name': SearchDatastore_Task, 'duration_secs': 0.008831} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.287713] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c11b1057-ff89-48ad-be73-ab578cdadc12 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.294194] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1023.294194] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e337b8-0592-0074-3755-205ca30ac29d" [ 1023.294194] env[61857]: _type = "Task" [ 1023.294194] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.310457] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e337b8-0592-0074-3755-205ca30ac29d, 'name': SearchDatastore_Task, 'duration_secs': 0.009508} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.310794] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.311137] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 91bb8c20-b7b6-4994-8a0a-b91935397036/91bb8c20-b7b6-4994-8a0a-b91935397036.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1023.311422] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bfa53354-3962-4e75-86c5-786e05190891 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.319409] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1023.319409] env[61857]: value = "task-951500" [ 1023.319409] env[61857]: _type = "Task" [ 1023.319409] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.327786] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951500, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.340327] env[61857]: DEBUG nova.network.neutron [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updated VIF entry in instance network info cache for port 83ff6e16-8b1b-4391-9ef1-57600502688c. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1023.340327] env[61857]: DEBUG nova.network.neutron [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance_info_cache with network_info: [{"id": "83ff6e16-8b1b-4391-9ef1-57600502688c", "address": "fa:16:3e:46:c4:d7", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83ff6e16-8b", "ovs_interfaceid": "83ff6e16-8b1b-4391-9ef1-57600502688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.595444] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951498, 'name': CreateVM_Task, 'duration_secs': 0.372312} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.595641] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1023.596403] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.596574] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.596955] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1023.597265] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-746626bf-5b28-41c4-bfe8-3d02304701ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.602490] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1023.602490] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523b8585-bbe7-412f-756e-202359def396" [ 1023.602490] env[61857]: _type = "Task" [ 1023.602490] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.611818] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523b8585-bbe7-412f-756e-202359def396, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.735934] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951499, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.829375] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951500, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476046} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.829680] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 91bb8c20-b7b6-4994-8a0a-b91935397036/91bb8c20-b7b6-4994-8a0a-b91935397036.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1023.829907] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.830187] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f8f86d8f-6249-46fc-a517-d5da8a3b8104 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.837537] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1023.837537] env[61857]: value = "task-951501" [ 1023.837537] env[61857]: _type = "Task" [ 1023.837537] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.842375] env[61857]: DEBUG oslo_concurrency.lockutils [req-01e6b68e-715f-4dfc-9718-ce700638e8d1 req-2a321382-4a68-46a5-b4a5-30ff0a6ee789 service nova] Releasing lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.847578] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951501, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.112745] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523b8585-bbe7-412f-756e-202359def396, 'name': SearchDatastore_Task, 'duration_secs': 0.05668} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.113101] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.113352] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.113599] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.113755] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.113938] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.114233] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c726c10-b909-4a97-ba8b-692918fb5034 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.122895] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.123191] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1024.124236] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86c496ed-db3c-4a88-b440-13a3d4ebe98c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.130647] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1024.130647] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ae314e-83cf-8a5c-4a18-660d25e6b4f8" [ 1024.130647] env[61857]: _type = "Task" [ 1024.130647] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.138329] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ae314e-83cf-8a5c-4a18-660d25e6b4f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.237191] env[61857]: DEBUG oslo_vmware.api [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951499, 'name': PowerOnVM_Task, 'duration_secs': 0.583266} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.237911] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1024.238150] env[61857]: INFO nova.compute.manager [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Took 8.03 seconds to spawn the instance on the hypervisor. [ 1024.238370] env[61857]: DEBUG nova.compute.manager [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.239184] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60088a13-aca3-47a8-b1ca-89be7f8a7885 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.254593] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d92635-4ab8-48e5-82de-84234abdced4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.261929] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fde67be-2bcf-424b-9f5c-4b7f52f509ef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.294562] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ce08a0-2402-4871-80be-400bb346a2f6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.302674] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e1b981-0aee-4baa-ab60-9550a2788df9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.317266] env[61857]: DEBUG nova.compute.provider_tree [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.348095] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951501, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077745} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.348445] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1024.349331] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374dc1da-93e6-4bfb-910e-18a3616463c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.373838] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 91bb8c20-b7b6-4994-8a0a-b91935397036/91bb8c20-b7b6-4994-8a0a-b91935397036.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.374246] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbc64ce4-9194-4e17-afc2-db35253f7897 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.394690] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1024.394690] env[61857]: value = "task-951502" [ 1024.394690] env[61857]: _type = "Task" [ 1024.394690] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.404750] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951502, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.643959] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ae314e-83cf-8a5c-4a18-660d25e6b4f8, 'name': SearchDatastore_Task, 'duration_secs': 0.008804} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.645189] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4583ab90-6640-4670-85eb-683c26e300b0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.652343] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1024.652343] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5210a5e2-8676-f3f9-ebea-f79352147c9e" [ 1024.652343] env[61857]: _type = "Task" [ 1024.652343] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.661938] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5210a5e2-8676-f3f9-ebea-f79352147c9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.760297] env[61857]: INFO nova.compute.manager [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Took 12.80 seconds to build instance. [ 1024.820080] env[61857]: DEBUG nova.scheduler.client.report [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.905133] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951502, 'name': ReconfigVM_Task, 'duration_secs': 0.503208} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.905379] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 91bb8c20-b7b6-4994-8a0a-b91935397036/91bb8c20-b7b6-4994-8a0a-b91935397036.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.906026] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-622114f6-d8f4-4ee8-83cf-c1ad041758fa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.912489] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1024.912489] env[61857]: value = "task-951503" [ 1024.912489] env[61857]: _type = "Task" [ 1024.912489] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.922374] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951503, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.162522] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5210a5e2-8676-f3f9-ebea-f79352147c9e, 'name': SearchDatastore_Task, 'duration_secs': 0.008981} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.162780] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.163563] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 04438fc0-b5a1-4e29-84d5-4d84d332c8e4/04438fc0-b5a1-4e29-84d5-4d84d332c8e4.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1025.163563] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e04cbec-2c29-44a5-b0c6-b4c2cfd4ca02 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.169767] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1025.169767] env[61857]: value = "task-951504" [ 1025.169767] env[61857]: _type = "Task" [ 1025.169767] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.178226] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.263080] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e36e7c6-4f2b-459f-8e2a-59036983b975 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.313s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.324809] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.324980] env[61857]: DEBUG nova.compute.manager [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1025.328157] env[61857]: DEBUG oslo_concurrency.lockutils [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.004s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.328675] env[61857]: DEBUG oslo_concurrency.lockutils [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.331046] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.347s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.331336] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.371041] env[61857]: INFO nova.scheduler.client.report [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted allocations for instance 94720bbd-a5db-45ca-9cfc-02c0f127c8bc [ 1025.425112] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951503, 'name': Rename_Task, 'duration_secs': 0.209799} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.425401] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1025.425696] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c47ab731-3efd-4c15-97e0-1f5f0876f41f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.434386] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1025.434386] env[61857]: value = "task-951505" [ 1025.434386] env[61857]: _type = "Task" [ 1025.434386] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.444625] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.526373] env[61857]: DEBUG oslo_vmware.rw_handles [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961971-94de-f848-cfbd-ec2a268ca36e/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1025.527566] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc38d7a-1a67-41b5-9aff-91ff3dfd2278 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.534541] env[61857]: DEBUG oslo_vmware.rw_handles [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961971-94de-f848-cfbd-ec2a268ca36e/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1025.534830] env[61857]: ERROR oslo_vmware.rw_handles [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961971-94de-f848-cfbd-ec2a268ca36e/disk-0.vmdk due to incomplete transfer. [ 1025.535061] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-62135576-c007-48d2-baae-dfdc3a19d291 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.542845] env[61857]: DEBUG oslo_vmware.rw_handles [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52961971-94de-f848-cfbd-ec2a268ca36e/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1025.543115] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Uploaded image 2382b3d3-d1ff-4aa0-90ba-7e846851d4d5 to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1025.553180] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1025.553591] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b45006be-9501-424b-893d-c70e8be0d8e9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.561583] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 1025.561583] env[61857]: value = "task-951506" [ 1025.561583] env[61857]: _type = "Task" [ 1025.561583] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.571057] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951506, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.681786] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951504, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.837065] env[61857]: DEBUG nova.compute.utils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1025.845037] env[61857]: DEBUG nova.compute.manager [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1025.845037] env[61857]: DEBUG nova.network.neutron [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1025.851649] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e2f54e4e-6a35-4204-b17f-951c4d78968f tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.345s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.851649] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 7.193s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.852432] env[61857]: INFO nova.compute.manager [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Unshelving [ 1025.883034] env[61857]: DEBUG oslo_concurrency.lockutils [None req-835d89c3-b0ea-4ef7-9f0c-5487cf597ac7 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "94720bbd-a5db-45ca-9cfc-02c0f127c8bc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.067s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.933378] env[61857]: DEBUG nova.compute.manager [req-082a747a-fcd7-465f-9d74-a1eae357520f req-704c7718-c027-4fc2-8165-92d2158dabab service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Received event network-changed-f3163567-eee7-4192-9332-fe1748882d93 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.933378] env[61857]: DEBUG nova.compute.manager [req-082a747a-fcd7-465f-9d74-a1eae357520f req-704c7718-c027-4fc2-8165-92d2158dabab service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Refreshing instance network info cache due to event network-changed-f3163567-eee7-4192-9332-fe1748882d93. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1025.933378] env[61857]: DEBUG oslo_concurrency.lockutils [req-082a747a-fcd7-465f-9d74-a1eae357520f req-704c7718-c027-4fc2-8165-92d2158dabab service nova] Acquiring lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.933378] env[61857]: DEBUG oslo_concurrency.lockutils [req-082a747a-fcd7-465f-9d74-a1eae357520f req-704c7718-c027-4fc2-8165-92d2158dabab service nova] Acquired lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.934632] env[61857]: DEBUG nova.network.neutron [req-082a747a-fcd7-465f-9d74-a1eae357520f req-704c7718-c027-4fc2-8165-92d2158dabab service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Refreshing network info cache for port f3163567-eee7-4192-9332-fe1748882d93 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1025.947067] env[61857]: DEBUG nova.policy [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9056d344ff614950a60983450a2a40ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5653ed6ab0a948ef89bb03cf7426dd10', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1025.955186] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951505, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.075077] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951506, 'name': Destroy_Task, 'duration_secs': 0.342769} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.075791] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Destroyed the VM [ 1026.076029] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1026.076424] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-45e7e278-ee1d-4462-a317-dda772a52eb0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.084233] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 1026.084233] env[61857]: value = "task-951507" [ 1026.084233] env[61857]: _type = "Task" [ 1026.084233] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.094243] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951507, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.181549] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549771} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.182491] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 04438fc0-b5a1-4e29-84d5-4d84d332c8e4/04438fc0-b5a1-4e29-84d5-4d84d332c8e4.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1026.183109] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1026.183578] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bbf902f6-3ecd-4de6-851d-6f8949e42ce3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.190725] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1026.190725] env[61857]: value = "task-951508" [ 1026.190725] env[61857]: _type = "Task" [ 1026.190725] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.199327] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951508, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.256512] env[61857]: DEBUG nova.network.neutron [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Successfully created port: a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1026.354266] env[61857]: DEBUG nova.compute.manager [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1026.456900] env[61857]: DEBUG oslo_vmware.api [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951505, 'name': PowerOnVM_Task, 'duration_secs': 0.98649} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.457242] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1026.457457] env[61857]: INFO nova.compute.manager [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Took 8.04 seconds to spawn the instance on the hypervisor. [ 1026.457640] env[61857]: DEBUG nova.compute.manager [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1026.458460] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138c0ec1-5581-4ab7-a439-0b7c27223982 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.598039] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951507, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.701033] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951508, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069362} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.701228] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1026.702209] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1264d05-394f-4e23-8a8e-da7fa277f608 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.726488] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 04438fc0-b5a1-4e29-84d5-4d84d332c8e4/04438fc0-b5a1-4e29-84d5-4d84d332c8e4.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.726902] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3272991-f464-4717-93a9-906d85cec2ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.746398] env[61857]: DEBUG nova.network.neutron [req-082a747a-fcd7-465f-9d74-a1eae357520f req-704c7718-c027-4fc2-8165-92d2158dabab service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updated VIF entry in instance network info cache for port f3163567-eee7-4192-9332-fe1748882d93. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1026.746781] env[61857]: DEBUG nova.network.neutron [req-082a747a-fcd7-465f-9d74-a1eae357520f req-704c7718-c027-4fc2-8165-92d2158dabab service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance_info_cache with network_info: [{"id": "f3163567-eee7-4192-9332-fe1748882d93", "address": "fa:16:3e:80:c4:35", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3163567-ee", "ovs_interfaceid": "f3163567-eee7-4192-9332-fe1748882d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.749224] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1026.749224] env[61857]: value = "task-951509" [ 1026.749224] env[61857]: _type = "Task" [ 1026.749224] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.758083] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951509, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.871108] env[61857]: DEBUG nova.compute.utils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1026.981966] env[61857]: INFO nova.compute.manager [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Took 14.60 seconds to build instance. [ 1027.100595] env[61857]: DEBUG oslo_vmware.api [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951507, 'name': RemoveSnapshot_Task, 'duration_secs': 0.542173} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.101130] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1027.101908] env[61857]: INFO nova.compute.manager [None req-6ff61432-673b-402c-8cd7-bfa770f449e9 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Took 13.18 seconds to snapshot the instance on the hypervisor. [ 1027.251308] env[61857]: DEBUG oslo_concurrency.lockutils [req-082a747a-fcd7-465f-9d74-a1eae357520f req-704c7718-c027-4fc2-8165-92d2158dabab service nova] Releasing lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.266414] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951509, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.374990] env[61857]: DEBUG nova.compute.manager [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1027.377994] env[61857]: INFO nova.virt.block_device [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Booting with volume f319926d-a413-4adf-b108-e133bfe9bd26 at /dev/sdb [ 1027.415706] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1027.415982] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1027.416169] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1027.416392] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1027.416564] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1027.416721] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1027.416935] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1027.417649] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1027.418033] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1027.418231] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1027.421652] env[61857]: DEBUG nova.virt.hardware [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.421652] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9913a0f9-7751-4cc9-ac68-94a8412e6a86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.423941] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5acc0384-49e9-439d-b68f-d65e7e0a5366 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.433484] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fd6876-2980-4f53-87df-b5a8656a0914 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.440664] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c40e84-d1cf-4a9d-ac19-0093ceb7856c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.481744] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7aa62b74-3939-49f8-9a20-98519e18718d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.483987] env[61857]: DEBUG oslo_concurrency.lockutils [None req-31752060-7367-443b-8a5a-3adabd248e2f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "91bb8c20-b7b6-4994-8a0a-b91935397036" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.109s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.490828] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0185b66d-1be6-4aea-a7a7-646f2bf2767d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.516786] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab416c05-7782-4919-9caa-93667f6807c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.523416] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb562f5-7666-4b2c-9ad4-4abfc46e459d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.537294] env[61857]: DEBUG nova.virt.block_device [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating existing volume attachment record: 2cf16f6b-b588-423d-8f11-2e0ac5647210 {{(pid=61857) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1027.545731] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquiring lock "9bb676fa-a53a-419a-bc22-1ed618d7191f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.545960] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lock "9bb676fa-a53a-419a-bc22-1ed618d7191f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.624490] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "5fff13ed-0914-4791-b518-ab0ab63fa52f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.625623] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "5fff13ed-0914-4791-b518-ab0ab63fa52f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.764143] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951509, 'name': ReconfigVM_Task, 'duration_secs': 0.711151} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.764528] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 04438fc0-b5a1-4e29-84d5-4d84d332c8e4/04438fc0-b5a1-4e29-84d5-4d84d332c8e4.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.765227] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-511cb670-8b0a-40c9-b67e-56a43127a93f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.771890] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1027.771890] env[61857]: value = "task-951511" [ 1027.771890] env[61857]: _type = "Task" [ 1027.771890] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.782393] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951511, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.961582] env[61857]: DEBUG nova.network.neutron [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Successfully updated port: a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1027.977862] env[61857]: DEBUG nova.compute.manager [req-04ca3268-ca8c-4e82-82ba-80cf883476aa req-f652b5b3-e9ad-4a0e-b6a5-302682d3ebd9 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received event network-vif-plugged-a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.978225] env[61857]: DEBUG oslo_concurrency.lockutils [req-04ca3268-ca8c-4e82-82ba-80cf883476aa req-f652b5b3-e9ad-4a0e-b6a5-302682d3ebd9 service nova] Acquiring lock "14844876-96ab-481d-980d-4133e89d7d4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.978569] env[61857]: DEBUG oslo_concurrency.lockutils [req-04ca3268-ca8c-4e82-82ba-80cf883476aa req-f652b5b3-e9ad-4a0e-b6a5-302682d3ebd9 service nova] Lock "14844876-96ab-481d-980d-4133e89d7d4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.978921] env[61857]: DEBUG oslo_concurrency.lockutils [req-04ca3268-ca8c-4e82-82ba-80cf883476aa req-f652b5b3-e9ad-4a0e-b6a5-302682d3ebd9 service nova] Lock "14844876-96ab-481d-980d-4133e89d7d4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.978921] env[61857]: DEBUG nova.compute.manager [req-04ca3268-ca8c-4e82-82ba-80cf883476aa req-f652b5b3-e9ad-4a0e-b6a5-302682d3ebd9 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] No waiting events found dispatching network-vif-plugged-a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1027.979074] env[61857]: WARNING nova.compute.manager [req-04ca3268-ca8c-4e82-82ba-80cf883476aa req-f652b5b3-e9ad-4a0e-b6a5-302682d3ebd9 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received unexpected event network-vif-plugged-a5cec797-c084-4a3f-828a-2624cd347348 for instance with vm_state building and task_state spawning. [ 1028.048873] env[61857]: DEBUG nova.compute.manager [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1028.127663] env[61857]: DEBUG nova.compute.manager [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1028.140578] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "91bb8c20-b7b6-4994-8a0a-b91935397036" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.140850] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "91bb8c20-b7b6-4994-8a0a-b91935397036" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.141035] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "91bb8c20-b7b6-4994-8a0a-b91935397036-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.141236] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "91bb8c20-b7b6-4994-8a0a-b91935397036-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.141406] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "91bb8c20-b7b6-4994-8a0a-b91935397036-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.143578] env[61857]: INFO nova.compute.manager [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Terminating instance [ 1028.145417] env[61857]: DEBUG nova.compute.manager [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1028.145618] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1028.146610] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6c2f20-3659-4901-911a-2643f13cf206 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.154667] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1028.154929] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae79a0a4-c8e2-476c-b5fc-150cce89c2ca {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.161917] env[61857]: DEBUG oslo_vmware.api [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1028.161917] env[61857]: value = "task-951514" [ 1028.161917] env[61857]: _type = "Task" [ 1028.161917] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.171521] env[61857]: DEBUG oslo_vmware.api [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.283031] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951511, 'name': Rename_Task, 'duration_secs': 0.137302} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.283031] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1028.283179] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b6916f5-73f2-4f5b-9273-403d995bb46e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.291137] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1028.291137] env[61857]: value = "task-951515" [ 1028.291137] env[61857]: _type = "Task" [ 1028.291137] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.298914] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951515, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.464672] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.465109] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.465109] env[61857]: DEBUG nova.network.neutron [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1028.571305] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquiring lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.571606] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.571815] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquiring lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.572027] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.572218] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.575048] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.575294] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.576744] env[61857]: INFO nova.compute.claims [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.581054] env[61857]: INFO nova.compute.manager [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Terminating instance [ 1028.581541] env[61857]: DEBUG nova.compute.manager [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1028.581737] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1028.582600] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeae4712-1f6a-422e-b956-246064d5b7bf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.591953] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1028.592214] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ca8f446-19a8-4e24-acae-8bb68d1f3480 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.599230] env[61857]: DEBUG oslo_vmware.api [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 1028.599230] env[61857]: value = "task-951516" [ 1028.599230] env[61857]: _type = "Task" [ 1028.599230] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.607658] env[61857]: DEBUG oslo_vmware.api [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.647106] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.671434] env[61857]: DEBUG oslo_vmware.api [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951514, 'name': PowerOffVM_Task, 'duration_secs': 0.252095} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.671717] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.671894] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1028.672195] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-febda494-9153-4e6a-bd39-9c1c2f8db33c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.742054] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1028.742278] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1028.742476] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleting the datastore file [datastore2] 91bb8c20-b7b6-4994-8a0a-b91935397036 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.742793] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-785c9631-a50b-4c29-a02d-34476b7b100c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.750171] env[61857]: DEBUG oslo_vmware.api [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1028.750171] env[61857]: value = "task-951518" [ 1028.750171] env[61857]: _type = "Task" [ 1028.750171] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.760849] env[61857]: DEBUG oslo_vmware.api [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.799698] env[61857]: DEBUG oslo_vmware.api [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951515, 'name': PowerOnVM_Task, 'duration_secs': 0.496304} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.799898] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1028.800126] env[61857]: INFO nova.compute.manager [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Took 8.12 seconds to spawn the instance on the hypervisor. [ 1028.800338] env[61857]: DEBUG nova.compute.manager [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.801151] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6101308f-03fb-4c66-aa6b-8b960337964f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.999202] env[61857]: DEBUG nova.network.neutron [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1029.109411] env[61857]: DEBUG oslo_vmware.api [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951516, 'name': PowerOffVM_Task, 'duration_secs': 0.274354} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.109688] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1029.109885] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1029.110141] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22c152b0-fa6f-404d-a65f-a7b7ef627a88 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.149797] env[61857]: DEBUG nova.network.neutron [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updating instance_info_cache with network_info: [{"id": "a5cec797-c084-4a3f-828a-2624cd347348", "address": "fa:16:3e:e7:99:f4", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5cec797-c0", "ovs_interfaceid": "a5cec797-c084-4a3f-828a-2624cd347348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.205526] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1029.205814] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1029.206016] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Deleting the datastore file [datastore1] e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.206291] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50acf949-d0e1-4e99-be92-962ccbffd6c5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.213463] env[61857]: DEBUG oslo_vmware.api [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for the task: (returnval){ [ 1029.213463] env[61857]: value = "task-951520" [ 1029.213463] env[61857]: _type = "Task" [ 1029.213463] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.221871] env[61857]: DEBUG oslo_vmware.api [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951520, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.259401] env[61857]: DEBUG oslo_vmware.api [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.387199} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.259678] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.259883] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1029.260081] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1029.260278] env[61857]: INFO nova.compute.manager [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1029.260538] env[61857]: DEBUG oslo.service.loopingcall [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.260760] env[61857]: DEBUG nova.compute.manager [-] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.260859] env[61857]: DEBUG nova.network.neutron [-] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1029.319604] env[61857]: INFO nova.compute.manager [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Took 16.44 seconds to build instance. [ 1029.653631] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.653974] env[61857]: DEBUG nova.compute.manager [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Instance network_info: |[{"id": "a5cec797-c084-4a3f-828a-2624cd347348", "address": "fa:16:3e:e7:99:f4", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5cec797-c0", "ovs_interfaceid": "a5cec797-c084-4a3f-828a-2624cd347348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1029.654613] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:99:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ffea58a-e043-4ad1-9ad7-45a61fdca001', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5cec797-c084-4a3f-828a-2624cd347348', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.665043] env[61857]: DEBUG oslo.service.loopingcall [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.669093] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1029.669521] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06e14ac7-e41b-4f4e-a729-52b14e1d47f7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.699637] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.699637] env[61857]: value = "task-951521" [ 1029.699637] env[61857]: _type = "Task" [ 1029.699637] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.712548] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951521, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.725045] env[61857]: DEBUG oslo_vmware.api [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Task: {'id': task-951520, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.37253} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.725340] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.725535] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1029.725718] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1029.725898] env[61857]: INFO nova.compute.manager [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1029.726295] env[61857]: DEBUG oslo.service.loopingcall [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.726521] env[61857]: DEBUG nova.compute.manager [-] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.726666] env[61857]: DEBUG nova.network.neutron [-] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1029.730037] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f48a7df-7445-40bd-b021-a8d022f97b53 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.737105] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5193c38f-d8c0-4d9d-896b-5dd9cf2786e3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.771668] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ec89c9-ac57-4bad-839b-e33b501bda30 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.781278] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90180af-c4e3-4150-aa41-ec41a6525d3c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.796449] env[61857]: DEBUG nova.compute.provider_tree [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.819808] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f25ce090-59f5-4c45-b28d-c8bd3de48da7 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.949s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.014645] env[61857]: DEBUG nova.compute.manager [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received event network-changed-a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1030.015086] env[61857]: DEBUG nova.compute.manager [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing instance network info cache due to event network-changed-a5cec797-c084-4a3f-828a-2624cd347348. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1030.015579] env[61857]: DEBUG oslo_concurrency.lockutils [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] Acquiring lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.015758] env[61857]: DEBUG oslo_concurrency.lockutils [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] Acquired lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.016067] env[61857]: DEBUG nova.network.neutron [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing network info cache for port a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1030.043962] env[61857]: DEBUG nova.compute.manager [req-84f18c34-f7ce-421e-9a60-5dcb5de45cc0 req-9f9a14ea-a6da-4f06-9fbc-a80746d4a4ff service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Received event network-changed-83ff6e16-8b1b-4391-9ef1-57600502688c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1030.044169] env[61857]: DEBUG nova.compute.manager [req-84f18c34-f7ce-421e-9a60-5dcb5de45cc0 req-9f9a14ea-a6da-4f06-9fbc-a80746d4a4ff service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Refreshing instance network info cache due to event network-changed-83ff6e16-8b1b-4391-9ef1-57600502688c. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1030.044612] env[61857]: DEBUG oslo_concurrency.lockutils [req-84f18c34-f7ce-421e-9a60-5dcb5de45cc0 req-9f9a14ea-a6da-4f06-9fbc-a80746d4a4ff service nova] Acquiring lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.044804] env[61857]: DEBUG oslo_concurrency.lockutils [req-84f18c34-f7ce-421e-9a60-5dcb5de45cc0 req-9f9a14ea-a6da-4f06-9fbc-a80746d4a4ff service nova] Acquired lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.044985] env[61857]: DEBUG nova.network.neutron [req-84f18c34-f7ce-421e-9a60-5dcb5de45cc0 req-9f9a14ea-a6da-4f06-9fbc-a80746d4a4ff service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Refreshing network info cache for port 83ff6e16-8b1b-4391-9ef1-57600502688c {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1030.069223] env[61857]: DEBUG nova.network.neutron [-] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.209442] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951521, 'name': CreateVM_Task, 'duration_secs': 0.349439} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.209674] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1030.210428] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.210608] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.210962] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1030.211273] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a29ec61-9525-4dc9-857c-3d96c70af2ea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.215897] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1030.215897] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e18ad2-3c42-58fc-75ca-03c21f653862" [ 1030.215897] env[61857]: _type = "Task" [ 1030.215897] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.223835] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e18ad2-3c42-58fc-75ca-03c21f653862, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.300166] env[61857]: DEBUG nova.scheduler.client.report [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.455024] env[61857]: DEBUG nova.network.neutron [-] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.571983] env[61857]: INFO nova.compute.manager [-] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Took 1.31 seconds to deallocate network for instance. [ 1030.726216] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52e18ad2-3c42-58fc-75ca-03c21f653862, 'name': SearchDatastore_Task, 'duration_secs': 0.008662} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.726563] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.726808] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1030.727057] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.727211] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.727411] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1030.727673] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f57c76d8-1add-4762-9fec-8a7c5669cc25 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.737606] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1030.737792] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1030.738632] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6621861-053c-4cb1-b00e-c95ff8959dc7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.744315] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1030.744315] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524ace0f-f3fe-f36c-03cc-40775f80aebf" [ 1030.744315] env[61857]: _type = "Task" [ 1030.744315] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.752430] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524ace0f-f3fe-f36c-03cc-40775f80aebf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.765644] env[61857]: DEBUG nova.network.neutron [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updated VIF entry in instance network info cache for port a5cec797-c084-4a3f-828a-2624cd347348. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1030.766024] env[61857]: DEBUG nova.network.neutron [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updating instance_info_cache with network_info: [{"id": "a5cec797-c084-4a3f-828a-2624cd347348", "address": "fa:16:3e:e7:99:f4", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5cec797-c0", "ovs_interfaceid": "a5cec797-c084-4a3f-828a-2624cd347348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.771736] env[61857]: DEBUG nova.network.neutron [req-84f18c34-f7ce-421e-9a60-5dcb5de45cc0 req-9f9a14ea-a6da-4f06-9fbc-a80746d4a4ff service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updated VIF entry in instance network info cache for port 83ff6e16-8b1b-4391-9ef1-57600502688c. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1030.772072] env[61857]: DEBUG nova.network.neutron [req-84f18c34-f7ce-421e-9a60-5dcb5de45cc0 req-9f9a14ea-a6da-4f06-9fbc-a80746d4a4ff service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance_info_cache with network_info: [{"id": "83ff6e16-8b1b-4391-9ef1-57600502688c", "address": "fa:16:3e:46:c4:d7", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83ff6e16-8b", "ovs_interfaceid": "83ff6e16-8b1b-4391-9ef1-57600502688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.805477] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.230s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.807870] env[61857]: DEBUG nova.compute.manager [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1030.808997] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.162s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.810408] env[61857]: INFO nova.compute.claims [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1030.958838] env[61857]: INFO nova.compute.manager [-] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Took 1.23 seconds to deallocate network for instance. [ 1031.079851] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.255089] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524ace0f-f3fe-f36c-03cc-40775f80aebf, 'name': SearchDatastore_Task, 'duration_secs': 0.008392} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.255788] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ed45210-1c5b-4e7d-bf83-3e1565a611ef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.261303] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1031.261303] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52015137-07a6-d6d8-b532-e22e1984a172" [ 1031.261303] env[61857]: _type = "Task" [ 1031.261303] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.269570] env[61857]: DEBUG oslo_concurrency.lockutils [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] Releasing lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.269818] env[61857]: DEBUG nova.compute.manager [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Received event network-vif-deleted-3a0cf2fa-0204-4f9e-8b78-2576d4e8524f {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1031.270000] env[61857]: INFO nova.compute.manager [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Neutron deleted interface 3a0cf2fa-0204-4f9e-8b78-2576d4e8524f; detaching it from the instance and deleting it from the info cache [ 1031.270195] env[61857]: DEBUG nova.network.neutron [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.271287] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52015137-07a6-d6d8-b532-e22e1984a172, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.273966] env[61857]: DEBUG oslo_concurrency.lockutils [req-84f18c34-f7ce-421e-9a60-5dcb5de45cc0 req-9f9a14ea-a6da-4f06-9fbc-a80746d4a4ff service nova] Releasing lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.310953] env[61857]: DEBUG nova.compute.utils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1031.312419] env[61857]: DEBUG nova.compute.manager [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1031.312613] env[61857]: DEBUG nova.network.neutron [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1031.355312] env[61857]: DEBUG nova.policy [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f5f8ea57881e4a248a739e54e543752c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1323346c811740bdb3bde112571e6708', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1031.465160] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.624783] env[61857]: DEBUG nova.network.neutron [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Successfully created port: cfb48455-4fd6-4614-b30b-0d8a27275c63 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1031.772886] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52015137-07a6-d6d8-b532-e22e1984a172, 'name': SearchDatastore_Task, 'duration_secs': 0.009028} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.773134] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5fce886e-ce6a-4250-94ad-31b8bbadb215 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.778021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.778021] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 14844876-96ab-481d-980d-4133e89d7d4d/14844876-96ab-481d-980d-4133e89d7d4d.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1031.778021] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39862408-f661-4119-96a9-a0150ba5f749 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.785940] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7e1905-24ac-4dcb-8894-6f5adf07d2c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.797617] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1031.797617] env[61857]: value = "task-951523" [ 1031.797617] env[61857]: _type = "Task" [ 1031.797617] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.807376] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951523, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.814190] env[61857]: DEBUG nova.compute.manager [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Detach interface failed, port_id=3a0cf2fa-0204-4f9e-8b78-2576d4e8524f, reason: Instance 91bb8c20-b7b6-4994-8a0a-b91935397036 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1031.814470] env[61857]: DEBUG nova.compute.manager [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Received event network-vif-deleted-77eb288b-3f38-448a-af1a-fada990f5ec4 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1031.814732] env[61857]: INFO nova.compute.manager [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Neutron deleted interface 77eb288b-3f38-448a-af1a-fada990f5ec4; detaching it from the instance and deleting it from the info cache [ 1031.814968] env[61857]: DEBUG nova.network.neutron [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.816577] env[61857]: DEBUG nova.compute.manager [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1031.987493] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0425f0-3c61-4299-96fd-091228649bf0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.999000] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a50909-fa38-41b4-ae86-634dfec2e434 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.032707] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7b41ef-c35f-4edf-a975-4f999dd49c44 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.042931] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b82950c-bb3f-4e86-ad08-4e19f8bf65db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.058935] env[61857]: DEBUG nova.compute.provider_tree [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.308162] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951523, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439374} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.308487] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 14844876-96ab-481d-980d-4133e89d7d4d/14844876-96ab-481d-980d-4133e89d7d4d.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1032.308716] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.308977] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-53a09e82-e41b-48a2-88bb-a361155fb327 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.315777] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1032.315777] env[61857]: value = "task-951524" [ 1032.315777] env[61857]: _type = "Task" [ 1032.315777] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.320736] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f60f32dd-4d78-4ad0-8539-4f564f75078b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.325298] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951524, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.333736] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073847cc-4879-40c2-aa5d-4ffe04607bfb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.359502] env[61857]: DEBUG nova.compute.manager [req-46be879e-dbe6-4377-8e5c-5fc558688806 req-8bd237f9-7c54-4956-8a6c-634482c9d6c6 service nova] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Detach interface failed, port_id=77eb288b-3f38-448a-af1a-fada990f5ec4, reason: Instance e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1032.561869] env[61857]: DEBUG nova.scheduler.client.report [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.825597] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951524, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06543} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.825897] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.826612] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da86c5eb-997a-47c0-b930-c4fd44537996 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.843118] env[61857]: DEBUG nova.compute.manager [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1032.855520] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 14844876-96ab-481d-980d-4133e89d7d4d/14844876-96ab-481d-980d-4133e89d7d4d.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.855520] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a40b2289-db09-4ea6-8a8d-d19c4a56a555 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.876019] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1032.876019] env[61857]: value = "task-951525" [ 1032.876019] env[61857]: _type = "Task" [ 1032.876019] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.883803] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951525, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.891697] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1032.891938] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1032.892379] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1032.892568] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1032.892755] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1032.892911] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1032.893140] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1032.893562] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1032.893562] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1032.893680] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1032.893819] env[61857]: DEBUG nova.virt.hardware [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1032.894986] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8c9d9d-fc07-4cf0-bc66-4efd46665607 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.902631] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-321d242c-1aa3-443a-a30e-4fb1984c4069 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.028087] env[61857]: DEBUG nova.compute.manager [req-fedc8376-50f1-4082-b346-dfd469802a55 req-0b31c62b-1fc7-4ad6-9246-6aa91c989bce service nova] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Received event network-vif-plugged-cfb48455-4fd6-4614-b30b-0d8a27275c63 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.028361] env[61857]: DEBUG oslo_concurrency.lockutils [req-fedc8376-50f1-4082-b346-dfd469802a55 req-0b31c62b-1fc7-4ad6-9246-6aa91c989bce service nova] Acquiring lock "9bb676fa-a53a-419a-bc22-1ed618d7191f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.028586] env[61857]: DEBUG oslo_concurrency.lockutils [req-fedc8376-50f1-4082-b346-dfd469802a55 req-0b31c62b-1fc7-4ad6-9246-6aa91c989bce service nova] Lock "9bb676fa-a53a-419a-bc22-1ed618d7191f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.028766] env[61857]: DEBUG oslo_concurrency.lockutils [req-fedc8376-50f1-4082-b346-dfd469802a55 req-0b31c62b-1fc7-4ad6-9246-6aa91c989bce service nova] Lock "9bb676fa-a53a-419a-bc22-1ed618d7191f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.028945] env[61857]: DEBUG nova.compute.manager [req-fedc8376-50f1-4082-b346-dfd469802a55 req-0b31c62b-1fc7-4ad6-9246-6aa91c989bce service nova] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] No waiting events found dispatching network-vif-plugged-cfb48455-4fd6-4614-b30b-0d8a27275c63 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1033.029454] env[61857]: WARNING nova.compute.manager [req-fedc8376-50f1-4082-b346-dfd469802a55 req-0b31c62b-1fc7-4ad6-9246-6aa91c989bce service nova] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Received unexpected event network-vif-plugged-cfb48455-4fd6-4614-b30b-0d8a27275c63 for instance with vm_state building and task_state spawning. [ 1033.066669] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.067238] env[61857]: DEBUG nova.compute.manager [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1033.069947] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.991s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.070192] env[61857]: DEBUG nova.objects.instance [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lazy-loading 'resources' on Instance uuid 91bb8c20-b7b6-4994-8a0a-b91935397036 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.143556] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.182412] env[61857]: DEBUG nova.network.neutron [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Successfully updated port: cfb48455-4fd6-4614-b30b-0d8a27275c63 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1033.385792] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951525, 'name': ReconfigVM_Task, 'duration_secs': 0.276753} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.386114] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 14844876-96ab-481d-980d-4133e89d7d4d/14844876-96ab-481d-980d-4133e89d7d4d.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.386736] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c174739c-08ea-4822-a2c1-9ddd4fcbc7a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.393786] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1033.393786] env[61857]: value = "task-951526" [ 1033.393786] env[61857]: _type = "Task" [ 1033.393786] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.401133] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951526, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.572973] env[61857]: DEBUG nova.compute.utils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1033.576816] env[61857]: DEBUG nova.compute.manager [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1033.576991] env[61857]: DEBUG nova.network.neutron [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1033.615388] env[61857]: DEBUG nova.policy [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74f5499cbed541deaf1905605e10158f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62bd314daba846cd9ed3dc994dd1034b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1033.685018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquiring lock "refresh_cache-9bb676fa-a53a-419a-bc22-1ed618d7191f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.685018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquired lock "refresh_cache-9bb676fa-a53a-419a-bc22-1ed618d7191f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.685018] env[61857]: DEBUG nova.network.neutron [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1033.700880] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e48a552b-ade2-4ab0-a84a-f29b97f1858c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.710572] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01550917-5412-43c3-9e38-02883695fd42 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.742291] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3661b803-6211-404b-8fd1-d19074e8f219 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.749649] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96de0fa-fbb3-4a97-ba8d-215a11d0c277 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.763474] env[61857]: DEBUG nova.compute.provider_tree [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.870194] env[61857]: DEBUG nova.network.neutron [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Successfully created port: 0cebc35e-e54f-47ec-80f8-39ed7194d942 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1033.904364] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951526, 'name': Rename_Task, 'duration_secs': 0.164468} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.904648] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1033.904897] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29ac68ea-38eb-473e-a1af-6aa5dc819d7c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.911511] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1033.911511] env[61857]: value = "task-951527" [ 1033.911511] env[61857]: _type = "Task" [ 1033.911511] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.918675] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951527, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.080041] env[61857]: DEBUG nova.compute.manager [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1034.230103] env[61857]: DEBUG nova.network.neutron [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1034.266270] env[61857]: DEBUG nova.scheduler.client.report [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.376569] env[61857]: DEBUG nova.network.neutron [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Updating instance_info_cache with network_info: [{"id": "cfb48455-4fd6-4614-b30b-0d8a27275c63", "address": "fa:16:3e:58:08:61", "network": {"id": "6092d7de-3fcb-4ee4-a651-dcffedfffbe1", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-650975045-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1323346c811740bdb3bde112571e6708", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8812601-ae67-4e0d-b9a2-710b86c53ac5", "external-id": "nsx-vlan-transportzone-85", "segmentation_id": 85, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfb48455-4f", "ovs_interfaceid": "cfb48455-4fd6-4614-b30b-0d8a27275c63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.421990] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951527, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.773370] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.703s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.775655] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.311s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.775890] env[61857]: DEBUG nova.objects.instance [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lazy-loading 'resources' on Instance uuid e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.795084] env[61857]: INFO nova.scheduler.client.report [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted allocations for instance 91bb8c20-b7b6-4994-8a0a-b91935397036 [ 1034.879359] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Releasing lock "refresh_cache-9bb676fa-a53a-419a-bc22-1ed618d7191f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.879875] env[61857]: DEBUG nova.compute.manager [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Instance network_info: |[{"id": "cfb48455-4fd6-4614-b30b-0d8a27275c63", "address": "fa:16:3e:58:08:61", "network": {"id": "6092d7de-3fcb-4ee4-a651-dcffedfffbe1", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-650975045-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1323346c811740bdb3bde112571e6708", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8812601-ae67-4e0d-b9a2-710b86c53ac5", "external-id": "nsx-vlan-transportzone-85", "segmentation_id": 85, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfb48455-4f", "ovs_interfaceid": "cfb48455-4fd6-4614-b30b-0d8a27275c63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1034.880157] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:08:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8812601-ae67-4e0d-b9a2-710b86c53ac5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cfb48455-4fd6-4614-b30b-0d8a27275c63', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1034.887473] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Creating folder: Project (1323346c811740bdb3bde112571e6708). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1034.887754] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9fff038-a052-4c91-9d58-1a81c49be02a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.899166] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Created folder: Project (1323346c811740bdb3bde112571e6708) in parent group-v214027. [ 1034.899355] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Creating folder: Instances. Parent ref: group-v214208. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1034.899581] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74c89fe7-8fd6-41e6-847a-70b5c4e1c173 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.909877] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Created folder: Instances in parent group-v214208. [ 1034.910114] env[61857]: DEBUG oslo.service.loopingcall [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.910347] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1034.910540] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbb8dbdf-1a37-4eb4-b6cd-952a5eb9ca54 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.932548] env[61857]: DEBUG oslo_vmware.api [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951527, 'name': PowerOnVM_Task, 'duration_secs': 0.558581} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.933636] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1034.933849] env[61857]: INFO nova.compute.manager [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Took 7.56 seconds to spawn the instance on the hypervisor. [ 1034.934055] env[61857]: DEBUG nova.compute.manager [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.934284] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1034.934284] env[61857]: value = "task-951530" [ 1034.934284] env[61857]: _type = "Task" [ 1034.934284] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.934921] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b686c3f3-7f05-4f69-b374-3a5bbeb42464 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.948635] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951530, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.061186] env[61857]: DEBUG nova.compute.manager [req-8b0b5fa8-2fe7-4983-9d57-003f976751c9 req-3ed7fbe5-2c2e-4d72-89f8-1c870d087aa4 service nova] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Received event network-changed-cfb48455-4fd6-4614-b30b-0d8a27275c63 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.061421] env[61857]: DEBUG nova.compute.manager [req-8b0b5fa8-2fe7-4983-9d57-003f976751c9 req-3ed7fbe5-2c2e-4d72-89f8-1c870d087aa4 service nova] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Refreshing instance network info cache due to event network-changed-cfb48455-4fd6-4614-b30b-0d8a27275c63. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1035.061669] env[61857]: DEBUG oslo_concurrency.lockutils [req-8b0b5fa8-2fe7-4983-9d57-003f976751c9 req-3ed7fbe5-2c2e-4d72-89f8-1c870d087aa4 service nova] Acquiring lock "refresh_cache-9bb676fa-a53a-419a-bc22-1ed618d7191f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.061827] env[61857]: DEBUG oslo_concurrency.lockutils [req-8b0b5fa8-2fe7-4983-9d57-003f976751c9 req-3ed7fbe5-2c2e-4d72-89f8-1c870d087aa4 service nova] Acquired lock "refresh_cache-9bb676fa-a53a-419a-bc22-1ed618d7191f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.062017] env[61857]: DEBUG nova.network.neutron [req-8b0b5fa8-2fe7-4983-9d57-003f976751c9 req-3ed7fbe5-2c2e-4d72-89f8-1c870d087aa4 service nova] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Refreshing network info cache for port cfb48455-4fd6-4614-b30b-0d8a27275c63 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1035.089237] env[61857]: DEBUG nova.compute.manager [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1035.119497] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1035.119774] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1035.119939] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1035.120148] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1035.120375] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1035.120705] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1035.120799] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1035.121023] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1035.121190] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1035.121378] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1035.121617] env[61857]: DEBUG nova.virt.hardware [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1035.122510] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483a3b24-a3e0-4bd2-b94f-1ed07a86e706 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.131119] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7d8bbb-7ff7-4aa2-95ca-a7855ef391fa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.304898] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1eaeb723-5d05-4ac8-8b3b-5c806e925da3 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "91bb8c20-b7b6-4994-8a0a-b91935397036" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.164s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.415677] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4cf298-b435-43a0-83cf-9251af30c632 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.424543] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e62a1f7-5d6e-48da-9c3b-0a229dc03769 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.464903] env[61857]: DEBUG nova.network.neutron [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Successfully updated port: 0cebc35e-e54f-47ec-80f8-39ed7194d942 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1035.469315] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294d895b-2bc3-42a6-a7b7-9c13d072f120 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.472852] env[61857]: INFO nova.compute.manager [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Took 21.00 seconds to build instance. [ 1035.481598] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff996eb7-5543-4477-85ed-880989fcd9ae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.485330] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951530, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.496677] env[61857]: DEBUG nova.compute.provider_tree [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.823166] env[61857]: DEBUG nova.network.neutron [req-8b0b5fa8-2fe7-4983-9d57-003f976751c9 req-3ed7fbe5-2c2e-4d72-89f8-1c870d087aa4 service nova] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Updated VIF entry in instance network info cache for port cfb48455-4fd6-4614-b30b-0d8a27275c63. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1035.823565] env[61857]: DEBUG nova.network.neutron [req-8b0b5fa8-2fe7-4983-9d57-003f976751c9 req-3ed7fbe5-2c2e-4d72-89f8-1c870d087aa4 service nova] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Updating instance_info_cache with network_info: [{"id": "cfb48455-4fd6-4614-b30b-0d8a27275c63", "address": "fa:16:3e:58:08:61", "network": {"id": "6092d7de-3fcb-4ee4-a651-dcffedfffbe1", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-650975045-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1323346c811740bdb3bde112571e6708", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8812601-ae67-4e0d-b9a2-710b86c53ac5", "external-id": "nsx-vlan-transportzone-85", "segmentation_id": 85, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfb48455-4f", "ovs_interfaceid": "cfb48455-4fd6-4614-b30b-0d8a27275c63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.967895] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951530, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.973919] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "refresh_cache-5fff13ed-0914-4791-b518-ab0ab63fa52f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.973919] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "refresh_cache-5fff13ed-0914-4791-b518-ab0ab63fa52f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.973919] env[61857]: DEBUG nova.network.neutron [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1035.975963] env[61857]: DEBUG nova.compute.manager [req-aa559a94-9f94-4093-85ed-c3e7f05070e4 req-b105daf7-15d5-4081-a093-74fe152a70d7 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received event network-changed-a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.976511] env[61857]: DEBUG nova.compute.manager [req-aa559a94-9f94-4093-85ed-c3e7f05070e4 req-b105daf7-15d5-4081-a093-74fe152a70d7 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing instance network info cache due to event network-changed-a5cec797-c084-4a3f-828a-2624cd347348. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1035.976511] env[61857]: DEBUG oslo_concurrency.lockutils [req-aa559a94-9f94-4093-85ed-c3e7f05070e4 req-b105daf7-15d5-4081-a093-74fe152a70d7 service nova] Acquiring lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.976511] env[61857]: DEBUG oslo_concurrency.lockutils [req-aa559a94-9f94-4093-85ed-c3e7f05070e4 req-b105daf7-15d5-4081-a093-74fe152a70d7 service nova] Acquired lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.976671] env[61857]: DEBUG nova.network.neutron [req-aa559a94-9f94-4093-85ed-c3e7f05070e4 req-b105daf7-15d5-4081-a093-74fe152a70d7 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing network info cache for port a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1035.978098] env[61857]: DEBUG oslo_concurrency.lockutils [None req-51edbedd-ba3b-42c3-89af-c4f5d139ff55 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14844876-96ab-481d-980d-4133e89d7d4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.512s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.999528] env[61857]: DEBUG nova.scheduler.client.report [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.318651] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.318916] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.328340] env[61857]: DEBUG oslo_concurrency.lockutils [req-8b0b5fa8-2fe7-4983-9d57-003f976751c9 req-3ed7fbe5-2c2e-4d72-89f8-1c870d087aa4 service nova] Releasing lock "refresh_cache-9bb676fa-a53a-419a-bc22-1ed618d7191f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.468496] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951530, 'name': CreateVM_Task, 'duration_secs': 1.307287} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.468678] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1036.469388] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.469564] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.469896] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1036.470164] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e3c2f17-ef31-4be3-9054-f554b67bf56a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.474488] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for the task: (returnval){ [ 1036.474488] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52a0104d-086a-1c66-f966-13ba3b106387" [ 1036.474488] env[61857]: _type = "Task" [ 1036.474488] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.487471] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a0104d-086a-1c66-f966-13ba3b106387, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.503915] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.728s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.506983] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.363s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.506983] env[61857]: DEBUG nova.objects.instance [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lazy-loading 'pci_requests' on Instance uuid a42e8c77-3cc3-4323-b1ee-2a95192910d5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.523531] env[61857]: DEBUG nova.network.neutron [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1036.526415] env[61857]: INFO nova.scheduler.client.report [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Deleted allocations for instance e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c [ 1036.662450] env[61857]: DEBUG nova.network.neutron [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Updating instance_info_cache with network_info: [{"id": "0cebc35e-e54f-47ec-80f8-39ed7194d942", "address": "fa:16:3e:d8:e9:ed", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cebc35e-e5", "ovs_interfaceid": "0cebc35e-e54f-47ec-80f8-39ed7194d942", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.706160] env[61857]: DEBUG nova.network.neutron [req-aa559a94-9f94-4093-85ed-c3e7f05070e4 req-b105daf7-15d5-4081-a093-74fe152a70d7 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updated VIF entry in instance network info cache for port a5cec797-c084-4a3f-828a-2624cd347348. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1036.706465] env[61857]: DEBUG nova.network.neutron [req-aa559a94-9f94-4093-85ed-c3e7f05070e4 req-b105daf7-15d5-4081-a093-74fe152a70d7 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updating instance_info_cache with network_info: [{"id": "a5cec797-c084-4a3f-828a-2624cd347348", "address": "fa:16:3e:e7:99:f4", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5cec797-c0", "ovs_interfaceid": "a5cec797-c084-4a3f-828a-2624cd347348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.821980] env[61857]: DEBUG nova.compute.manager [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1036.895039] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "14df609d-678c-4713-8c9c-f80fabb61c79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.895297] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14df609d-678c-4713-8c9c-f80fabb61c79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.984441] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52a0104d-086a-1c66-f966-13ba3b106387, 'name': SearchDatastore_Task, 'duration_secs': 0.024983} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.984742] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.985252] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1036.985252] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.985367] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.986624] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1036.986624] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9dde0cb8-0e08-40b2-955a-8fc4e8d67e4b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.995232] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1036.995327] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1036.995982] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2edbf51b-c678-44ba-9db4-ae92e6f486e8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.001154] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for the task: (returnval){ [ 1037.001154] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525756ed-ca87-9430-39cf-38af38d6c09d" [ 1037.001154] env[61857]: _type = "Task" [ 1037.001154] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.008854] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525756ed-ca87-9430-39cf-38af38d6c09d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.013524] env[61857]: DEBUG nova.objects.instance [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lazy-loading 'numa_topology' on Instance uuid a42e8c77-3cc3-4323-b1ee-2a95192910d5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.039352] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e73923a9-1595-402c-863d-028cd0aa5173 tempest-ImagesOneServerTestJSON-1019871312 tempest-ImagesOneServerTestJSON-1019871312-project-member] Lock "e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.468s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.092656] env[61857]: DEBUG nova.compute.manager [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Received event network-vif-plugged-0cebc35e-e54f-47ec-80f8-39ed7194d942 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1037.092913] env[61857]: DEBUG oslo_concurrency.lockutils [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] Acquiring lock "5fff13ed-0914-4791-b518-ab0ab63fa52f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.093339] env[61857]: DEBUG oslo_concurrency.lockutils [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] Lock "5fff13ed-0914-4791-b518-ab0ab63fa52f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.093546] env[61857]: DEBUG oslo_concurrency.lockutils [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] Lock "5fff13ed-0914-4791-b518-ab0ab63fa52f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.093731] env[61857]: DEBUG nova.compute.manager [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] No waiting events found dispatching network-vif-plugged-0cebc35e-e54f-47ec-80f8-39ed7194d942 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1037.093909] env[61857]: WARNING nova.compute.manager [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Received unexpected event network-vif-plugged-0cebc35e-e54f-47ec-80f8-39ed7194d942 for instance with vm_state building and task_state spawning. [ 1037.094089] env[61857]: DEBUG nova.compute.manager [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Received event network-changed-0cebc35e-e54f-47ec-80f8-39ed7194d942 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1037.094260] env[61857]: DEBUG nova.compute.manager [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Refreshing instance network info cache due to event network-changed-0cebc35e-e54f-47ec-80f8-39ed7194d942. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1037.094487] env[61857]: DEBUG oslo_concurrency.lockutils [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] Acquiring lock "refresh_cache-5fff13ed-0914-4791-b518-ab0ab63fa52f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.165093] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "refresh_cache-5fff13ed-0914-4791-b518-ab0ab63fa52f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.165465] env[61857]: DEBUG nova.compute.manager [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Instance network_info: |[{"id": "0cebc35e-e54f-47ec-80f8-39ed7194d942", "address": "fa:16:3e:d8:e9:ed", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cebc35e-e5", "ovs_interfaceid": "0cebc35e-e54f-47ec-80f8-39ed7194d942", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1037.165794] env[61857]: DEBUG oslo_concurrency.lockutils [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] Acquired lock "refresh_cache-5fff13ed-0914-4791-b518-ab0ab63fa52f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.166028] env[61857]: DEBUG nova.network.neutron [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Refreshing network info cache for port 0cebc35e-e54f-47ec-80f8-39ed7194d942 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1037.167278] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:e9:ed', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0cebc35e-e54f-47ec-80f8-39ed7194d942', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1037.176343] env[61857]: DEBUG oslo.service.loopingcall [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.179510] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1037.180033] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7345eb6-5215-496c-bb82-ec6efedd36ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.200472] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1037.200472] env[61857]: value = "task-951531" [ 1037.200472] env[61857]: _type = "Task" [ 1037.200472] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.208236] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951531, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.210081] env[61857]: DEBUG oslo_concurrency.lockutils [req-aa559a94-9f94-4093-85ed-c3e7f05070e4 req-b105daf7-15d5-4081-a093-74fe152a70d7 service nova] Releasing lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.358593] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.399308] env[61857]: DEBUG nova.compute.manager [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1037.417836] env[61857]: DEBUG nova.network.neutron [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Updated VIF entry in instance network info cache for port 0cebc35e-e54f-47ec-80f8-39ed7194d942. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1037.418331] env[61857]: DEBUG nova.network.neutron [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Updating instance_info_cache with network_info: [{"id": "0cebc35e-e54f-47ec-80f8-39ed7194d942", "address": "fa:16:3e:d8:e9:ed", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cebc35e-e5", "ovs_interfaceid": "0cebc35e-e54f-47ec-80f8-39ed7194d942", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.512282] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525756ed-ca87-9430-39cf-38af38d6c09d, 'name': SearchDatastore_Task, 'duration_secs': 0.026168} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.513087] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3af5afd8-5c59-424b-9d29-17c426ec3211 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.515617] env[61857]: INFO nova.compute.claims [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.521194] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for the task: (returnval){ [ 1037.521194] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52f2fad6-b70b-d1cf-8421-af21f1227be8" [ 1037.521194] env[61857]: _type = "Task" [ 1037.521194] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.529046] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f2fad6-b70b-d1cf-8421-af21f1227be8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.710999] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951531, 'name': CreateVM_Task, 'duration_secs': 0.313705} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.711188] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1037.711870] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.712053] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.712399] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.712643] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-687eec60-bd8d-48e6-b63e-5bd16fc44436 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.717017] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1037.717017] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525a530a-2aec-5b9a-2f0f-3be66a333e40" [ 1037.717017] env[61857]: _type = "Task" [ 1037.717017] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.724325] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525a530a-2aec-5b9a-2f0f-3be66a333e40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.921329] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.921645] env[61857]: DEBUG oslo_concurrency.lockutils [req-5e68018b-e08a-4f55-9893-cc827de0b7d7 req-5f83fe86-678f-48e9-827b-612b78bd6c62 service nova] Releasing lock "refresh_cache-5fff13ed-0914-4791-b518-ab0ab63fa52f" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.033117] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f2fad6-b70b-d1cf-8421-af21f1227be8, 'name': SearchDatastore_Task, 'duration_secs': 0.029544} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.033770] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.033770] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 9bb676fa-a53a-419a-bc22-1ed618d7191f/9bb676fa-a53a-419a-bc22-1ed618d7191f.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1038.034510] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ddd3ac03-aa30-4fc0-8571-92cfe0c39f1d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.042940] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for the task: (returnval){ [ 1038.042940] env[61857]: value = "task-951532" [ 1038.042940] env[61857]: _type = "Task" [ 1038.042940] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.052440] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951532, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.168834] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e38ffc-f8de-419a-aaf8-56a0054d1ba7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.176658] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf73bcc9-459d-4def-ab8d-c8ea45b5b970 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.210240] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5365509-680b-44d5-a88b-08a0b0de987e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.218911] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cde22fb-27db-47aa-8794-c33a69ca2839 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.230944] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525a530a-2aec-5b9a-2f0f-3be66a333e40, 'name': SearchDatastore_Task, 'duration_secs': 0.038849} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.238543] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.238808] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1038.239112] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.239234] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.239731] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1038.240191] env[61857]: DEBUG nova.compute.provider_tree [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.241163] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0f8e441-ffb6-4fc2-8836-010eb12778ca {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.249710] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1038.249865] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1038.251734] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d0aa9ef-d06b-4230-ab05-f96d38c53271 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.256916] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1038.256916] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52366d8b-4aa4-a1d6-b5fe-436371295f75" [ 1038.256916] env[61857]: _type = "Task" [ 1038.256916] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.265632] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52366d8b-4aa4-a1d6-b5fe-436371295f75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.556811] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951532, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.744695] env[61857]: DEBUG nova.scheduler.client.report [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.768126] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52366d8b-4aa4-a1d6-b5fe-436371295f75, 'name': SearchDatastore_Task, 'duration_secs': 0.009186} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.768962] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f0e8525-98e4-43e5-bc3d-ea3c6558035a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.773716] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1038.773716] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]522f2a1f-860e-1de4-f449-20aa67ad14a9" [ 1038.773716] env[61857]: _type = "Task" [ 1038.773716] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.780667] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522f2a1f-860e-1de4-f449-20aa67ad14a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.055190] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951532, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67169} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.055548] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 9bb676fa-a53a-419a-bc22-1ed618d7191f/9bb676fa-a53a-419a-bc22-1ed618d7191f.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1039.055662] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.055928] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7982ab6e-22b9-4584-a1ef-0c5c7e43d7c8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.062187] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for the task: (returnval){ [ 1039.062187] env[61857]: value = "task-951533" [ 1039.062187] env[61857]: _type = "Task" [ 1039.062187] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.071571] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951533, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.250478] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.744s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.252731] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.894s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.254245] env[61857]: INFO nova.compute.claims [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.283775] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522f2a1f-860e-1de4-f449-20aa67ad14a9, 'name': SearchDatastore_Task, 'duration_secs': 0.050889} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.284049] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.284385] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 5fff13ed-0914-4791-b518-ab0ab63fa52f/5fff13ed-0914-4791-b518-ab0ab63fa52f.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1039.284590] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04fd85d1-d3a6-431c-add8-f148174befe7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.291384] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1039.291384] env[61857]: value = "task-951534" [ 1039.291384] env[61857]: _type = "Task" [ 1039.291384] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.300205] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951534, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.301122] env[61857]: INFO nova.network.neutron [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating port 507983f7-c4c5-4605-887e-c3a1936e8f3a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1039.574084] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951533, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.258726} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.574369] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1039.575174] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d156302d-e7fd-4d09-bfb3-cd1c0ca5168f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.596788] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 9bb676fa-a53a-419a-bc22-1ed618d7191f/9bb676fa-a53a-419a-bc22-1ed618d7191f.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.597077] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b4343f9-c5e8-42da-9545-784eb0eabe16 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.616556] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for the task: (returnval){ [ 1039.616556] env[61857]: value = "task-951535" [ 1039.616556] env[61857]: _type = "Task" [ 1039.616556] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.624386] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951535, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.801540] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951534, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.126183] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951535, 'name': ReconfigVM_Task, 'duration_secs': 0.277422} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.126634] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 9bb676fa-a53a-419a-bc22-1ed618d7191f/9bb676fa-a53a-419a-bc22-1ed618d7191f.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.127094] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-16c5f7aa-a156-438c-bb45-10621570dbdf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.132833] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for the task: (returnval){ [ 1040.132833] env[61857]: value = "task-951536" [ 1040.132833] env[61857]: _type = "Task" [ 1040.132833] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.139863] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951536, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.302672] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951534, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.376914] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a92144-28fa-41ba-9461-e49e0d942ec8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.384558] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9405c18b-2613-4694-84e0-ab85f58fdd02 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.416805] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e3557a2-f7af-41f0-818d-74bbbaf354be {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.424476] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9b92af-b474-42e3-a313-c736730a8ed1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.438443] env[61857]: DEBUG nova.compute.provider_tree [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.644249] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951536, 'name': Rename_Task, 'duration_secs': 0.127373} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.644572] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1040.644844] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-84da267e-e38f-4a58-a56a-78898430d559 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.652050] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for the task: (returnval){ [ 1040.652050] env[61857]: value = "task-951537" [ 1040.652050] env[61857]: _type = "Task" [ 1040.652050] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.662013] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951537, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.802834] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951534, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.435625} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.803166] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 5fff13ed-0914-4791-b518-ab0ab63fa52f/5fff13ed-0914-4791-b518-ab0ab63fa52f.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1040.803396] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1040.803643] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a31296cd-f1cc-479c-b678-517aba01b16f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.809956] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1040.809956] env[61857]: value = "task-951538" [ 1040.809956] env[61857]: _type = "Task" [ 1040.809956] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.817273] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951538, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.941720] env[61857]: DEBUG nova.scheduler.client.report [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.161453] env[61857]: DEBUG oslo_vmware.api [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951537, 'name': PowerOnVM_Task, 'duration_secs': 0.420325} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.161748] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1041.161927] env[61857]: INFO nova.compute.manager [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Took 8.32 seconds to spawn the instance on the hypervisor. [ 1041.162122] env[61857]: DEBUG nova.compute.manager [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1041.162863] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2939bb9f-4d14-46ba-8a1c-a7e46c46597d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.319534] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951538, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060622} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.319818] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1041.320713] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96367f87-c61c-4ada-84ef-1a9a9b156a7a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.343015] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 5fff13ed-0914-4791-b518-ab0ab63fa52f/5fff13ed-0914-4791-b518-ab0ab63fa52f.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.343292] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f47e2178-80a0-4c88-ad18-1dd55c1b5744 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.364961] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1041.364961] env[61857]: value = "task-951539" [ 1041.364961] env[61857]: _type = "Task" [ 1041.364961] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.374138] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951539, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.425247] env[61857]: DEBUG nova.compute.manager [req-4c784bb9-3472-4a7c-aa1e-7868f6244257 req-6fd3d9c3-687d-4502-af0f-c503ca0eaa25 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received event network-vif-plugged-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1041.425247] env[61857]: DEBUG oslo_concurrency.lockutils [req-4c784bb9-3472-4a7c-aa1e-7868f6244257 req-6fd3d9c3-687d-4502-af0f-c503ca0eaa25 service nova] Acquiring lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.425247] env[61857]: DEBUG oslo_concurrency.lockutils [req-4c784bb9-3472-4a7c-aa1e-7868f6244257 req-6fd3d9c3-687d-4502-af0f-c503ca0eaa25 service nova] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.425247] env[61857]: DEBUG oslo_concurrency.lockutils [req-4c784bb9-3472-4a7c-aa1e-7868f6244257 req-6fd3d9c3-687d-4502-af0f-c503ca0eaa25 service nova] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.425247] env[61857]: DEBUG nova.compute.manager [req-4c784bb9-3472-4a7c-aa1e-7868f6244257 req-6fd3d9c3-687d-4502-af0f-c503ca0eaa25 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] No waiting events found dispatching network-vif-plugged-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1041.425247] env[61857]: WARNING nova.compute.manager [req-4c784bb9-3472-4a7c-aa1e-7868f6244257 req-6fd3d9c3-687d-4502-af0f-c503ca0eaa25 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received unexpected event network-vif-plugged-507983f7-c4c5-4605-887e-c3a1936e8f3a for instance with vm_state shelved_offloaded and task_state spawning. [ 1041.449018] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.194s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.449018] env[61857]: DEBUG nova.compute.manager [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1041.449978] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.529s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.453878] env[61857]: INFO nova.compute.claims [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1041.513137] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.513137] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.513137] env[61857]: DEBUG nova.network.neutron [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1041.679529] env[61857]: INFO nova.compute.manager [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Took 13.12 seconds to build instance. [ 1041.877685] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951539, 'name': ReconfigVM_Task, 'duration_secs': 0.312585} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.877685] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 5fff13ed-0914-4791-b518-ab0ab63fa52f/5fff13ed-0914-4791-b518-ab0ab63fa52f.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.878089] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d149c6cb-a9e2-4aa1-9810-cef87a5b1a7b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.887796] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1041.887796] env[61857]: value = "task-951540" [ 1041.887796] env[61857]: _type = "Task" [ 1041.887796] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.895446] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951540, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.961103] env[61857]: DEBUG nova.compute.utils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.962803] env[61857]: DEBUG nova.compute.manager [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1041.963192] env[61857]: DEBUG nova.network.neutron [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1042.004443] env[61857]: DEBUG nova.policy [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2ef7e845007475a8a19cf3f520c3f51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1acf476aabc4166bc8505a3442367c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1042.181795] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a0faec7c-486d-4f4d-9b7d-ae1af84d2f32 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lock "9bb676fa-a53a-419a-bc22-1ed618d7191f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.636s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.287653] env[61857]: DEBUG nova.network.neutron [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Successfully created port: 63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1042.290555] env[61857]: DEBUG nova.network.neutron [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating instance_info_cache with network_info: [{"id": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "address": "fa:16:3e:52:2f:3f", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507983f7-c4", "ovs_interfaceid": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.398626] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951540, 'name': Rename_Task, 'duration_secs': 0.134231} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.398763] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1042.399027] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4d99bc7-4263-4a2f-9f0d-3014313d127d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.406496] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1042.406496] env[61857]: value = "task-951541" [ 1042.406496] env[61857]: _type = "Task" [ 1042.406496] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.418393] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951541, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.469265] env[61857]: DEBUG nova.compute.manager [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1042.605965] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22de5ed-8580-448c-86ec-d582c1579466 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.613855] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1041142-88f0-4fd5-be6d-5e6a82a77af9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.650832] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ffcd483-28d7-45b4-8c04-94a560ce3049 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.658641] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b4fca2-755a-4d93-9d51-706a7f46c08f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.672208] env[61857]: DEBUG nova.compute.provider_tree [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.793715] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.823765] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='356ddc8ed89c3407916f08032c5f33ac',container_format='bare',created_at=2024-09-18T21:46:01Z,direct_url=,disk_format='vmdk',id=6f2ff642-ffbc-477c-83d9-8ce6daf6b818,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-2025326519-shelved',owner='35ae3038a282412e963845e221e87f67',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-09-18T21:46:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.824114] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.824318] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.824556] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.824764] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.824972] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.825271] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.825477] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.825668] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.825869] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.826188] env[61857]: DEBUG nova.virt.hardware [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.827106] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4400843-b00c-4305-ba78-360a0791bbe8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.835504] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89cfff34-0007-43e1-b84e-7cef043b770a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.849049] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:2f:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '507983f7-c4c5-4605-887e-c3a1936e8f3a', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.856350] env[61857]: DEBUG oslo.service.loopingcall [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.856608] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1042.856819] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-593b957a-184c-46ca-af40-d4f9f9c9195c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.876580] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.876580] env[61857]: value = "task-951542" [ 1042.876580] env[61857]: _type = "Task" [ 1042.876580] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.886689] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951542, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.915418] env[61857]: DEBUG oslo_vmware.api [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951541, 'name': PowerOnVM_Task, 'duration_secs': 0.43311} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.915760] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1042.915984] env[61857]: INFO nova.compute.manager [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Took 7.83 seconds to spawn the instance on the hypervisor. [ 1042.916307] env[61857]: DEBUG nova.compute.manager [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.916990] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753e2907-d5c5-49ec-8499-7ec791d681ce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.985910] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquiring lock "9bb676fa-a53a-419a-bc22-1ed618d7191f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.986204] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lock "9bb676fa-a53a-419a-bc22-1ed618d7191f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.986447] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquiring lock "9bb676fa-a53a-419a-bc22-1ed618d7191f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.986655] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lock "9bb676fa-a53a-419a-bc22-1ed618d7191f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.986831] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lock "9bb676fa-a53a-419a-bc22-1ed618d7191f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.989163] env[61857]: INFO nova.compute.manager [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Terminating instance [ 1042.991377] env[61857]: DEBUG nova.compute.manager [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1042.991604] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1042.992549] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51cba43-aaae-490e-936c-38cf20af638a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.004217] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1043.004497] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc8af085-d4a7-47c2-b331-8a1e69db77e7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.011399] env[61857]: DEBUG oslo_vmware.api [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for the task: (returnval){ [ 1043.011399] env[61857]: value = "task-951543" [ 1043.011399] env[61857]: _type = "Task" [ 1043.011399] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.019821] env[61857]: DEBUG oslo_vmware.api [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951543, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.175335] env[61857]: DEBUG nova.scheduler.client.report [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.389163] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951542, 'name': CreateVM_Task, 'duration_secs': 0.311269} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.389494] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1043.390303] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.390546] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.391094] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.391441] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-934cb3a8-2695-4101-8681-906f5e79f2af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.396958] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1043.396958] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52cb9dad-35f9-2887-7d73-89ed6ee594ca" [ 1043.396958] env[61857]: _type = "Task" [ 1043.396958] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.411299] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52cb9dad-35f9-2887-7d73-89ed6ee594ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.436860] env[61857]: INFO nova.compute.manager [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Took 14.80 seconds to build instance. [ 1043.453034] env[61857]: DEBUG nova.compute.manager [req-66071412-4c40-4e1e-afb3-d33ffdb08e00 req-5bdf3ee5-5896-4456-a6fc-393a1ebde8d8 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received event network-changed-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1043.453340] env[61857]: DEBUG nova.compute.manager [req-66071412-4c40-4e1e-afb3-d33ffdb08e00 req-5bdf3ee5-5896-4456-a6fc-393a1ebde8d8 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Refreshing instance network info cache due to event network-changed-507983f7-c4c5-4605-887e-c3a1936e8f3a. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1043.453673] env[61857]: DEBUG oslo_concurrency.lockutils [req-66071412-4c40-4e1e-afb3-d33ffdb08e00 req-5bdf3ee5-5896-4456-a6fc-393a1ebde8d8 service nova] Acquiring lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.453904] env[61857]: DEBUG oslo_concurrency.lockutils [req-66071412-4c40-4e1e-afb3-d33ffdb08e00 req-5bdf3ee5-5896-4456-a6fc-393a1ebde8d8 service nova] Acquired lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.454175] env[61857]: DEBUG nova.network.neutron [req-66071412-4c40-4e1e-afb3-d33ffdb08e00 req-5bdf3ee5-5896-4456-a6fc-393a1ebde8d8 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Refreshing network info cache for port 507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1043.484369] env[61857]: DEBUG nova.compute.manager [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1043.506979] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1043.507344] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1043.507595] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.507881] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1043.508134] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.508393] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1043.508701] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1043.508951] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1043.509230] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1043.509598] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1043.509887] env[61857]: DEBUG nova.virt.hardware [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1043.511090] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605a2e87-3789-4ea6-ab0b-dbf5c7972463 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.529339] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01cb725b-dbef-4b74-ab07-def44d77c166 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.534733] env[61857]: DEBUG oslo_vmware.api [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951543, 'name': PowerOffVM_Task, 'duration_secs': 0.18519} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.535424] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1043.535686] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1043.536410] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-30f36633-d9f3-4499-a19a-66d8a1754364 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.614683] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1043.614935] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1043.615140] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Deleting the datastore file [datastore1] 9bb676fa-a53a-419a-bc22-1ed618d7191f {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.615420] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb748fa6-b875-48a1-8f41-a046e583dd5e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.622106] env[61857]: DEBUG oslo_vmware.api [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for the task: (returnval){ [ 1043.622106] env[61857]: value = "task-951545" [ 1043.622106] env[61857]: _type = "Task" [ 1043.622106] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.630154] env[61857]: DEBUG oslo_vmware.api [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951545, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.681558] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.231s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.682162] env[61857]: DEBUG nova.compute.manager [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1043.908033] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.908033] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Processing image 6f2ff642-ffbc-477c-83d9-8ce6daf6b818 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.908033] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818/6f2ff642-ffbc-477c-83d9-8ce6daf6b818.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.908033] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818/6f2ff642-ffbc-477c-83d9-8ce6daf6b818.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.908278] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.908515] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1200874-3762-47d4-a1c8-8e1f802840dc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.925467] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.925612] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1043.926326] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d46aa9d-bf57-499d-997a-17eb7739fc19 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.931185] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1043.931185] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52487918-3dde-7917-c9c1-e03c2850ab19" [ 1043.931185] env[61857]: _type = "Task" [ 1043.931185] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.938887] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c5e7842e-5da6-4ee3-9d1f-2ef8ba8ff7df tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "5fff13ed-0914-4791-b518-ab0ab63fa52f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.314s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.939072] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52487918-3dde-7917-c9c1-e03c2850ab19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.959411] env[61857]: DEBUG nova.network.neutron [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Successfully updated port: 63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1044.131585] env[61857]: DEBUG oslo_vmware.api [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Task: {'id': task-951545, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.373399} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.131855] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.132063] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1044.132242] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1044.132422] env[61857]: INFO nova.compute.manager [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1044.132684] env[61857]: DEBUG oslo.service.loopingcall [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.132893] env[61857]: DEBUG nova.compute.manager [-] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1044.132986] env[61857]: DEBUG nova.network.neutron [-] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1044.176599] env[61857]: DEBUG nova.network.neutron [req-66071412-4c40-4e1e-afb3-d33ffdb08e00 req-5bdf3ee5-5896-4456-a6fc-393a1ebde8d8 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updated VIF entry in instance network info cache for port 507983f7-c4c5-4605-887e-c3a1936e8f3a. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1044.176599] env[61857]: DEBUG nova.network.neutron [req-66071412-4c40-4e1e-afb3-d33ffdb08e00 req-5bdf3ee5-5896-4456-a6fc-393a1ebde8d8 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating instance_info_cache with network_info: [{"id": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "address": "fa:16:3e:52:2f:3f", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507983f7-c4", "ovs_interfaceid": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.188376] env[61857]: DEBUG nova.compute.utils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1044.190527] env[61857]: DEBUG nova.compute.manager [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1044.190527] env[61857]: DEBUG nova.network.neutron [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1044.231451] env[61857]: DEBUG nova.policy [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9056d344ff614950a60983450a2a40ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5653ed6ab0a948ef89bb03cf7426dd10', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1044.441458] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Preparing fetch location {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1044.441793] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Fetch image to [datastore2] OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e/OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e.vmdk {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1044.441925] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Downloading stream optimized image 6f2ff642-ffbc-477c-83d9-8ce6daf6b818 to [datastore2] OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e/OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e.vmdk on the data store datastore2 as vApp {{(pid=61857) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1044.442137] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Downloading image file data 6f2ff642-ffbc-477c-83d9-8ce6daf6b818 to the ESX as VM named 'OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e' {{(pid=61857) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1044.462121] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "refresh_cache-7ab9803b-8baa-4c49-b8df-61cd3b64b216" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.462238] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "refresh_cache-7ab9803b-8baa-4c49-b8df-61cd3b64b216" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.462413] env[61857]: DEBUG nova.network.neutron [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1044.523507] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8653824-9872-45fa-814a-8425f665545a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.528665] env[61857]: DEBUG oslo_vmware.rw_handles [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1044.528665] env[61857]: value = "resgroup-9" [ 1044.528665] env[61857]: _type = "ResourcePool" [ 1044.528665] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1044.528983] env[61857]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-dbc64dc0-aa7a-46f9-a525-19982bb3a5cd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.547402] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ae574917-2385-4914-a28b-96260d4c1327 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Suspending the VM {{(pid=61857) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1044.548356] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-729c95ea-5534-4315-a49f-bcd478fa2c70 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.553010] env[61857]: DEBUG oslo_vmware.rw_handles [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lease: (returnval){ [ 1044.553010] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520298a7-b0ed-f685-e1ef-63783532870e" [ 1044.553010] env[61857]: _type = "HttpNfcLease" [ 1044.553010] env[61857]: } obtained for vApp import into resource pool (val){ [ 1044.553010] env[61857]: value = "resgroup-9" [ 1044.553010] env[61857]: _type = "ResourcePool" [ 1044.553010] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1044.553301] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the lease: (returnval){ [ 1044.553301] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520298a7-b0ed-f685-e1ef-63783532870e" [ 1044.553301] env[61857]: _type = "HttpNfcLease" [ 1044.553301] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1044.557589] env[61857]: DEBUG oslo_vmware.api [None req-ae574917-2385-4914-a28b-96260d4c1327 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1044.557589] env[61857]: value = "task-951547" [ 1044.557589] env[61857]: _type = "Task" [ 1044.557589] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.562474] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1044.562474] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520298a7-b0ed-f685-e1ef-63783532870e" [ 1044.562474] env[61857]: _type = "HttpNfcLease" [ 1044.562474] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1044.569508] env[61857]: DEBUG oslo_vmware.api [None req-ae574917-2385-4914-a28b-96260d4c1327 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951547, 'name': SuspendVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.602783] env[61857]: DEBUG nova.network.neutron [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Successfully created port: a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1044.678972] env[61857]: DEBUG oslo_concurrency.lockutils [req-66071412-4c40-4e1e-afb3-d33ffdb08e00 req-5bdf3ee5-5896-4456-a6fc-393a1ebde8d8 service nova] Releasing lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.693130] env[61857]: DEBUG nova.compute.manager [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1044.853684] env[61857]: DEBUG nova.network.neutron [-] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.002512] env[61857]: DEBUG nova.network.neutron [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1045.071096] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1045.071096] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520298a7-b0ed-f685-e1ef-63783532870e" [ 1045.071096] env[61857]: _type = "HttpNfcLease" [ 1045.071096] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1045.075101] env[61857]: DEBUG oslo_vmware.rw_handles [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1045.075101] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520298a7-b0ed-f685-e1ef-63783532870e" [ 1045.075101] env[61857]: _type = "HttpNfcLease" [ 1045.075101] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1045.075399] env[61857]: DEBUG oslo_vmware.api [None req-ae574917-2385-4914-a28b-96260d4c1327 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951547, 'name': SuspendVM_Task} progress is 62%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.077228] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f23f08-196d-4320-a5a8-f88a1932a841 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.087456] env[61857]: DEBUG oslo_vmware.rw_handles [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e66b9-f464-8b14-83d3-12fff1e5300c/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1045.087749] env[61857]: DEBUG oslo_vmware.rw_handles [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e66b9-f464-8b14-83d3-12fff1e5300c/disk-0.vmdk. {{(pid=61857) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1045.156882] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-967bc146-16e7-419a-b617-bce1179182d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.209519] env[61857]: DEBUG nova.network.neutron [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Updating instance_info_cache with network_info: [{"id": "63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31", "address": "fa:16:3e:13:3e:10", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f7bcf4-84", "ovs_interfaceid": "63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.357067] env[61857]: INFO nova.compute.manager [-] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Took 1.22 seconds to deallocate network for instance. [ 1045.486354] env[61857]: DEBUG nova.compute.manager [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Received event network-vif-plugged-63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1045.486683] env[61857]: DEBUG oslo_concurrency.lockutils [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] Acquiring lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.486895] env[61857]: DEBUG oslo_concurrency.lockutils [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] Lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.487116] env[61857]: DEBUG oslo_concurrency.lockutils [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] Lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.487390] env[61857]: DEBUG nova.compute.manager [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] No waiting events found dispatching network-vif-plugged-63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1045.487625] env[61857]: WARNING nova.compute.manager [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Received unexpected event network-vif-plugged-63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31 for instance with vm_state building and task_state spawning. [ 1045.487828] env[61857]: DEBUG nova.compute.manager [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Received event network-changed-63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1045.488064] env[61857]: DEBUG nova.compute.manager [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Refreshing instance network info cache due to event network-changed-63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1045.488317] env[61857]: DEBUG oslo_concurrency.lockutils [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] Acquiring lock "refresh_cache-7ab9803b-8baa-4c49-b8df-61cd3b64b216" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.575018] env[61857]: DEBUG oslo_vmware.api [None req-ae574917-2385-4914-a28b-96260d4c1327 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951547, 'name': SuspendVM_Task, 'duration_secs': 0.755772} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.575328] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ae574917-2385-4914-a28b-96260d4c1327 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Suspended the VM {{(pid=61857) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1045.575516] env[61857]: DEBUG nova.compute.manager [None req-ae574917-2385-4914-a28b-96260d4c1327 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1045.576333] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2dbbe5-a94b-4a9a-85a7-a5e72e960f0d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.705110] env[61857]: DEBUG nova.compute.manager [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1045.712487] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "refresh_cache-7ab9803b-8baa-4c49-b8df-61cd3b64b216" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.713123] env[61857]: DEBUG nova.compute.manager [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Instance network_info: |[{"id": "63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31", "address": "fa:16:3e:13:3e:10", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f7bcf4-84", "ovs_interfaceid": "63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1045.715335] env[61857]: DEBUG oslo_concurrency.lockutils [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] Acquired lock "refresh_cache-7ab9803b-8baa-4c49-b8df-61cd3b64b216" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.715551] env[61857]: DEBUG nova.network.neutron [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Refreshing network info cache for port 63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1045.717026] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:3e:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1045.725988] env[61857]: DEBUG oslo.service.loopingcall [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.734060] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1045.736906] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-afdaad65-8e15-49f3-8a8f-882b27d8b8e4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.758030] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1045.758030] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1045.758324] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.758439] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1045.758821] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.758821] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1045.758938] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1045.759478] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1045.759478] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1045.759478] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1045.759772] env[61857]: DEBUG nova.virt.hardware [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.761104] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ddb505-950e-4bf0-99a2-c9dfacf2ee79 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.776099] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbfbccc-8857-4c9f-8fcd-2f2471c7fac2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.781451] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1045.781451] env[61857]: value = "task-951548" [ 1045.781451] env[61857]: _type = "Task" [ 1045.781451] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.814372] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951548, 'name': CreateVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.868358] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.869308] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.869308] env[61857]: DEBUG nova.objects.instance [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lazy-loading 'resources' on Instance uuid 9bb676fa-a53a-419a-bc22-1ed618d7191f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.002874] env[61857]: DEBUG nova.network.neutron [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Updated VIF entry in instance network info cache for port 63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.003356] env[61857]: DEBUG nova.network.neutron [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Updating instance_info_cache with network_info: [{"id": "63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31", "address": "fa:16:3e:13:3e:10", "network": {"id": "3febde63-1ca9-4e39-9040-49605d83cbf8", "bridge": "br-int", "label": "tempest-ServersTestJSON-556177852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1acf476aabc4166bc8505a3442367c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f7bcf4-84", "ovs_interfaceid": "63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.241464] env[61857]: DEBUG nova.compute.manager [req-a62f5279-b50f-4b8a-a0f1-1e727b2e5dc3 req-1c6138c3-90de-4724-8912-73f906f9fec5 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Received event network-vif-plugged-a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.241719] env[61857]: DEBUG oslo_concurrency.lockutils [req-a62f5279-b50f-4b8a-a0f1-1e727b2e5dc3 req-1c6138c3-90de-4724-8912-73f906f9fec5 service nova] Acquiring lock "14df609d-678c-4713-8c9c-f80fabb61c79-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.241973] env[61857]: DEBUG oslo_concurrency.lockutils [req-a62f5279-b50f-4b8a-a0f1-1e727b2e5dc3 req-1c6138c3-90de-4724-8912-73f906f9fec5 service nova] Lock "14df609d-678c-4713-8c9c-f80fabb61c79-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.242728] env[61857]: DEBUG oslo_concurrency.lockutils [req-a62f5279-b50f-4b8a-a0f1-1e727b2e5dc3 req-1c6138c3-90de-4724-8912-73f906f9fec5 service nova] Lock "14df609d-678c-4713-8c9c-f80fabb61c79-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.242936] env[61857]: DEBUG nova.compute.manager [req-a62f5279-b50f-4b8a-a0f1-1e727b2e5dc3 req-1c6138c3-90de-4724-8912-73f906f9fec5 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] No waiting events found dispatching network-vif-plugged-a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1046.243138] env[61857]: WARNING nova.compute.manager [req-a62f5279-b50f-4b8a-a0f1-1e727b2e5dc3 req-1c6138c3-90de-4724-8912-73f906f9fec5 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Received unexpected event network-vif-plugged-a8571bda-024e-4d91-9417-6b5f64325f87 for instance with vm_state building and task_state spawning. [ 1046.295672] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951548, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.331595] env[61857]: DEBUG nova.network.neutron [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Successfully updated port: a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.500873] env[61857]: DEBUG oslo_vmware.rw_handles [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Completed reading data from the image iterator. {{(pid=61857) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1046.501288] env[61857]: DEBUG oslo_vmware.rw_handles [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e66b9-f464-8b14-83d3-12fff1e5300c/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1046.502133] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba26424e-8ace-4b07-83b9-62afb7552e92 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.506976] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed34e6f-8415-4147-8e97-a4e0d6812208 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.510389] env[61857]: DEBUG oslo_concurrency.lockutils [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] Releasing lock "refresh_cache-7ab9803b-8baa-4c49-b8df-61cd3b64b216" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.510709] env[61857]: DEBUG nova.compute.manager [req-6dc5aa65-4ff1-4e97-9b71-13ec35ae5729 req-d649718c-70e0-41fe-97d7-5b22b406236b service nova] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Received event network-vif-deleted-cfb48455-4fd6-4614-b30b-0d8a27275c63 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.514196] env[61857]: DEBUG oslo_vmware.rw_handles [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e66b9-f464-8b14-83d3-12fff1e5300c/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1046.514430] env[61857]: DEBUG oslo_vmware.rw_handles [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e66b9-f464-8b14-83d3-12fff1e5300c/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1046.516323] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-19cbff1c-bbf6-4995-9109-1eedc1bd261d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.518638] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0418c938-20de-486d-87c0-fa0285b3016f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.557152] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65075178-b69e-4b9f-bccf-951104330cb2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.568665] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e1efd7-819c-4dfc-9c8f-7cde974bf450 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.585275] env[61857]: DEBUG nova.compute.provider_tree [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.705070] env[61857]: DEBUG oslo_vmware.rw_handles [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e66b9-f464-8b14-83d3-12fff1e5300c/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1046.705866] env[61857]: INFO nova.virt.vmwareapi.images [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Downloaded image file data 6f2ff642-ffbc-477c-83d9-8ce6daf6b818 [ 1046.706318] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cafaab7-36f7-46f6-9168-5ea56e006b00 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.722664] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9fbc235-03b7-469a-b46b-8b549322e9c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.751096] env[61857]: INFO nova.virt.vmwareapi.images [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] The imported VM was unregistered [ 1046.753536] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Caching image {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1046.753770] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating directory with path [datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818 {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.754079] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-207debf5-00e0-4a39-8d04-b3579e690abc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.765085] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Created directory with path [datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818 {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.765286] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e/OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e.vmdk to [datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818/6f2ff642-ffbc-477c-83d9-8ce6daf6b818.vmdk. {{(pid=61857) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1046.765543] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-b4214cd3-42ff-435b-86b8-728c3542f7c3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.772698] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1046.772698] env[61857]: value = "task-951550" [ 1046.772698] env[61857]: _type = "Task" [ 1046.772698] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.780420] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951550, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.794125] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951548, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.834448] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.834671] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.834862] env[61857]: DEBUG nova.network.neutron [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1047.063459] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "5fff13ed-0914-4791-b518-ab0ab63fa52f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.063734] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "5fff13ed-0914-4791-b518-ab0ab63fa52f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.063967] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "5fff13ed-0914-4791-b518-ab0ab63fa52f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.064208] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "5fff13ed-0914-4791-b518-ab0ab63fa52f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.064386] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "5fff13ed-0914-4791-b518-ab0ab63fa52f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.067439] env[61857]: INFO nova.compute.manager [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Terminating instance [ 1047.069540] env[61857]: DEBUG nova.compute.manager [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1047.069746] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1047.070637] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1044b3a4-f6b2-4c36-86c7-295a0ef2cb0d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.078823] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1047.079097] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc1fa1bc-179e-4eb4-8fec-9daa7bb0b77b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.088388] env[61857]: DEBUG nova.scheduler.client.report [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.154757] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1047.155030] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1047.155230] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleting the datastore file [datastore1] 5fff13ed-0914-4791-b518-ab0ab63fa52f {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1047.155513] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-173ea017-3be0-42f3-a867-1f6af8418a3e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.164029] env[61857]: DEBUG oslo_vmware.api [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1047.164029] env[61857]: value = "task-951552" [ 1047.164029] env[61857]: _type = "Task" [ 1047.164029] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.174779] env[61857]: DEBUG oslo_vmware.api [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951552, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.283336] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951550, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.296653] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951548, 'name': CreateVM_Task, 'duration_secs': 1.48686} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.296827] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1047.297562] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.297743] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.298091] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.298369] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb6e0427-3d61-4a78-a861-088accbc655e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.304440] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1047.304440] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520e7382-33c1-a71d-3902-838dd2c09433" [ 1047.304440] env[61857]: _type = "Task" [ 1047.304440] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.312913] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]520e7382-33c1-a71d-3902-838dd2c09433, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.368909] env[61857]: DEBUG nova.network.neutron [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1047.520021] env[61857]: DEBUG nova.network.neutron [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updating instance_info_cache with network_info: [{"id": "a8571bda-024e-4d91-9417-6b5f64325f87", "address": "fa:16:3e:e8:b5:c8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8571bda-02", "ovs_interfaceid": "a8571bda-024e-4d91-9417-6b5f64325f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.593909] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.725s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.613082] env[61857]: INFO nova.scheduler.client.report [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Deleted allocations for instance 9bb676fa-a53a-419a-bc22-1ed618d7191f [ 1047.676229] env[61857]: DEBUG oslo_vmware.api [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951552, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257431} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.676229] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1047.676229] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1047.676229] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1047.676229] env[61857]: INFO nova.compute.manager [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1047.676229] env[61857]: DEBUG oslo.service.loopingcall [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.676229] env[61857]: DEBUG nova.compute.manager [-] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1047.676667] env[61857]: DEBUG nova.network.neutron [-] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1047.787153] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951550, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.816348] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]520e7382-33c1-a71d-3902-838dd2c09433, 'name': SearchDatastore_Task, 'duration_secs': 0.073587} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.816696] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.816940] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.817205] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.817407] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.817806] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.818126] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-43a7c0c5-081a-40c8-875c-cd598b4b8ada {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.834942] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.835285] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1047.836108] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9618846e-e800-4bb3-af18-a6bc9b163611 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.843547] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1047.843547] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]529ccb3c-4a1c-1a2b-f1e8-851b9a5ab11c" [ 1047.843547] env[61857]: _type = "Task" [ 1047.843547] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.854068] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529ccb3c-4a1c-1a2b-f1e8-851b9a5ab11c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.023327] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.023327] env[61857]: DEBUG nova.compute.manager [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Instance network_info: |[{"id": "a8571bda-024e-4d91-9417-6b5f64325f87", "address": "fa:16:3e:e8:b5:c8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8571bda-02", "ovs_interfaceid": "a8571bda-024e-4d91-9417-6b5f64325f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1048.023572] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:b5:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ffea58a-e043-4ad1-9ad7-45a61fdca001', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8571bda-024e-4d91-9417-6b5f64325f87', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1048.031288] env[61857]: DEBUG oslo.service.loopingcall [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.032507] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1048.032507] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e72c487f-5e57-47cc-98b1-339c054cc024 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.056076] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1048.056076] env[61857]: value = "task-951553" [ 1048.056076] env[61857]: _type = "Task" [ 1048.056076] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.065821] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951553, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.120856] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0afc3036-3f55-4224-999f-ca8a05146b89 tempest-ServerTagsTestJSON-11495882 tempest-ServerTagsTestJSON-11495882-project-member] Lock "9bb676fa-a53a-419a-bc22-1ed618d7191f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.134s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.287568] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951550, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.295196] env[61857]: DEBUG nova.compute.manager [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Received event network-changed-a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.295471] env[61857]: DEBUG nova.compute.manager [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Refreshing instance network info cache due to event network-changed-a8571bda-024e-4d91-9417-6b5f64325f87. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1048.295630] env[61857]: DEBUG oslo_concurrency.lockutils [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] Acquiring lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.295773] env[61857]: DEBUG oslo_concurrency.lockutils [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] Acquired lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.295940] env[61857]: DEBUG nova.network.neutron [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Refreshing network info cache for port a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1048.354021] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529ccb3c-4a1c-1a2b-f1e8-851b9a5ab11c, 'name': SearchDatastore_Task, 'duration_secs': 0.069027} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.354881] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8da8fe3e-4866-4657-90a1-13054395a17c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.361440] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1048.361440] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5231cfc1-5a57-5f3c-b420-c797fe30a721" [ 1048.361440] env[61857]: _type = "Task" [ 1048.361440] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.371276] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5231cfc1-5a57-5f3c-b420-c797fe30a721, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.473674] env[61857]: DEBUG nova.network.neutron [-] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.568672] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951553, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.786482] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951550, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.873758] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5231cfc1-5a57-5f3c-b420-c797fe30a721, 'name': SearchDatastore_Task, 'duration_secs': 0.075678} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.874172] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.874412] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 7ab9803b-8baa-4c49-b8df-61cd3b64b216/7ab9803b-8baa-4c49-b8df-61cd3b64b216.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1048.874784] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ece450a6-6270-4f24-ac69-fe1ae4b89787 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.883453] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1048.883453] env[61857]: value = "task-951554" [ 1048.883453] env[61857]: _type = "Task" [ 1048.883453] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.895635] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951554, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.976700] env[61857]: INFO nova.compute.manager [-] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Took 1.30 seconds to deallocate network for instance. [ 1049.071813] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951553, 'name': CreateVM_Task} progress is 25%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.290229] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951550, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.396615] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951554, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.486942] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.487256] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.487507] env[61857]: DEBUG nova.objects.instance [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lazy-loading 'resources' on Instance uuid 5fff13ed-0914-4791-b518-ab0ab63fa52f {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.502319] env[61857]: DEBUG nova.network.neutron [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updated VIF entry in instance network info cache for port a8571bda-024e-4d91-9417-6b5f64325f87. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1049.502926] env[61857]: DEBUG nova.network.neutron [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updating instance_info_cache with network_info: [{"id": "a8571bda-024e-4d91-9417-6b5f64325f87", "address": "fa:16:3e:e8:b5:c8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8571bda-02", "ovs_interfaceid": "a8571bda-024e-4d91-9417-6b5f64325f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.569533] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951553, 'name': CreateVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.787538] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951550, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.774907} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.787906] env[61857]: INFO nova.virt.vmwareapi.ds_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e/OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e.vmdk to [datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818/6f2ff642-ffbc-477c-83d9-8ce6daf6b818.vmdk. [ 1049.788127] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Cleaning up location [datastore2] OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1049.788352] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_e567722a-53de-45a3-933a-9a64eaa2b80e {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.788648] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7abf1fad-890b-4d5c-903a-ed9967c3cd0c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.795908] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1049.795908] env[61857]: value = "task-951555" [ 1049.795908] env[61857]: _type = "Task" [ 1049.795908] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.804879] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951555, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.894207] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951554, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.005649] env[61857]: DEBUG oslo_concurrency.lockutils [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] Releasing lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.005948] env[61857]: DEBUG nova.compute.manager [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Received event network-vif-deleted-0cebc35e-e54f-47ec-80f8-39ed7194d942 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1050.006149] env[61857]: INFO nova.compute.manager [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Neutron deleted interface 0cebc35e-e54f-47ec-80f8-39ed7194d942; detaching it from the instance and deleting it from the info cache [ 1050.006333] env[61857]: DEBUG nova.network.neutron [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.071330] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951553, 'name': CreateVM_Task, 'duration_secs': 1.524344} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.071651] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1050.072236] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.072407] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.072726] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1050.072981] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-564f52de-8224-4a70-9989-6ed6a16a81b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.079405] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1050.079405] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5256310a-87fc-6598-34e8-9c6dfed93a49" [ 1050.079405] env[61857]: _type = "Task" [ 1050.079405] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.087191] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5256310a-87fc-6598-34e8-9c6dfed93a49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.099620] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb458db-4ca7-46ff-a618-9c3cc6b4ad2e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.106244] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4890da47-b0a0-4862-a016-3e01f02b96b6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.138833] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6b20b3-28fa-4fd2-af6a-9b12f67a7910 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.145922] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f186cd-b161-4805-a327-3e1ece37150a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.158909] env[61857]: DEBUG nova.compute.provider_tree [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.306582] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951555, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122434} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.306866] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.307062] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818/6f2ff642-ffbc-477c-83d9-8ce6daf6b818.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.307323] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818/6f2ff642-ffbc-477c-83d9-8ce6daf6b818.vmdk to [datastore2] a42e8c77-3cc3-4323-b1ee-2a95192910d5/a42e8c77-3cc3-4323-b1ee-2a95192910d5.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1050.307593] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dacdc1d7-4d89-4915-9995-57a55cd3bd63 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.314524] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1050.314524] env[61857]: value = "task-951556" [ 1050.314524] env[61857]: _type = "Task" [ 1050.314524] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.322472] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951556, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.393692] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951554, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.041663} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.393928] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 7ab9803b-8baa-4c49-b8df-61cd3b64b216/7ab9803b-8baa-4c49-b8df-61cd3b64b216.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1050.394169] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1050.394452] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-852e32c3-c9e7-4f0d-9094-3fce72d316de {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.400586] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1050.400586] env[61857]: value = "task-951557" [ 1050.400586] env[61857]: _type = "Task" [ 1050.400586] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.407796] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951557, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.511464] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-48965355-aa39-4a9e-82c1-9a66967b5f9c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.521518] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9bae08-5645-4f6e-bc48-852f77027770 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.549320] env[61857]: DEBUG nova.compute.manager [req-571f73e5-c2ae-40e8-9c64-f9409673ed93 req-ea5deaca-4ce6-47d6-afc2-e010565b84db service nova] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Detach interface failed, port_id=0cebc35e-e54f-47ec-80f8-39ed7194d942, reason: Instance 5fff13ed-0914-4791-b518-ab0ab63fa52f could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1050.589696] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5256310a-87fc-6598-34e8-9c6dfed93a49, 'name': SearchDatastore_Task, 'duration_secs': 0.008621} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.590081] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.590330] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1050.590607] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.590773] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.590990] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1050.591304] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dac39762-6ec9-4d7b-ae88-e766e7079e18 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.605252] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1050.605562] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1050.606363] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bf1f398-c1c2-42b6-aba9-afd825d39f95 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.614433] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1050.614433] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]522a0662-3f44-7b9b-adf0-59de3527d7c1" [ 1050.614433] env[61857]: _type = "Task" [ 1050.614433] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.622683] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522a0662-3f44-7b9b-adf0-59de3527d7c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.662019] env[61857]: DEBUG nova.scheduler.client.report [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.825291] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951556, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.911024] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951557, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.152907} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.911274] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.912086] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc002a6-53a1-46f7-b84b-e2e276cf606c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.934821] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 7ab9803b-8baa-4c49-b8df-61cd3b64b216/7ab9803b-8baa-4c49-b8df-61cd3b64b216.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.935222] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee398956-9f02-43fb-aca4-c7ef28f9f267 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.955888] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1050.955888] env[61857]: value = "task-951558" [ 1050.955888] env[61857]: _type = "Task" [ 1050.955888] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.964834] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951558, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.126048] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522a0662-3f44-7b9b-adf0-59de3527d7c1, 'name': SearchDatastore_Task, 'duration_secs': 0.062893} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.126583] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c1dafd7-ec76-400e-8ad0-951feeadc8f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.132044] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1051.132044] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ee0bef-2e4c-735b-b468-26d777005254" [ 1051.132044] env[61857]: _type = "Task" [ 1051.132044] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.140622] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ee0bef-2e4c-735b-b468-26d777005254, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.166858] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.679s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.192503] env[61857]: INFO nova.scheduler.client.report [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted allocations for instance 5fff13ed-0914-4791-b518-ab0ab63fa52f [ 1051.326463] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951556, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.465637] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.642235] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ee0bef-2e4c-735b-b468-26d777005254, 'name': SearchDatastore_Task, 'duration_secs': 0.071993} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.642637] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.642802] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 14df609d-678c-4713-8c9c-f80fabb61c79/14df609d-678c-4713-8c9c-f80fabb61c79.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1051.643093] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-943dca0b-63ee-427b-91e6-5ad7d58af1cd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.649890] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1051.649890] env[61857]: value = "task-951559" [ 1051.649890] env[61857]: _type = "Task" [ 1051.649890] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.700977] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fc6a39e4-4e72-4c30-8888-cd656c868736 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "5fff13ed-0914-4791-b518-ab0ab63fa52f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.637s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.827942] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951556, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.966573] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.160213] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951559, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.327043] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951556, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.467196] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.659530] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951559, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.829139] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951556, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.333552} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.829452] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6f2ff642-ffbc-477c-83d9-8ce6daf6b818/6f2ff642-ffbc-477c-83d9-8ce6daf6b818.vmdk to [datastore2] a42e8c77-3cc3-4323-b1ee-2a95192910d5/a42e8c77-3cc3-4323-b1ee-2a95192910d5.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1052.830427] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a525e61-c533-4b75-b349-a636ee5fde46 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.853398] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] a42e8c77-3cc3-4323-b1ee-2a95192910d5/a42e8c77-3cc3-4323-b1ee-2a95192910d5.vmdk or device None with type streamOptimized {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1052.853735] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3f5ffe7-11a0-4fd4-abff-0395eb02f15b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.873534] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1052.873534] env[61857]: value = "task-951560" [ 1052.873534] env[61857]: _type = "Task" [ 1052.873534] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.881701] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951560, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.968138] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951558, 'name': ReconfigVM_Task, 'duration_secs': 1.858215} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.968509] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 7ab9803b-8baa-4c49-b8df-61cd3b64b216/7ab9803b-8baa-4c49-b8df-61cd3b64b216.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1052.969207] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-878254f4-144b-4710-89bc-3ebb4e9b1b1f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.977063] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1052.977063] env[61857]: value = "task-951561" [ 1052.977063] env[61857]: _type = "Task" [ 1052.977063] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.985813] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951561, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.001014] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "01708c16-43f8-408a-ae6b-82a1ad34eafa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.001118] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.160816] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951559, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.350599} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.161050] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 14df609d-678c-4713-8c9c-f80fabb61c79/14df609d-678c-4713-8c9c-f80fabb61c79.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1053.161305] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1053.161580] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ce94220-5580-4e4f-a30f-a0f00169aa31 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.169282] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1053.169282] env[61857]: value = "task-951562" [ 1053.169282] env[61857]: _type = "Task" [ 1053.169282] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.175632] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951562, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.384587] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951560, 'name': ReconfigVM_Task, 'duration_secs': 0.339384} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.384903] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Reconfigured VM instance instance-00000059 to attach disk [datastore2] a42e8c77-3cc3-4323-b1ee-2a95192910d5/a42e8c77-3cc3-4323-b1ee-2a95192910d5.vmdk or device None with type streamOptimized {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.386251] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_secret_uuid': None, 'boot_index': 0, 'guest_format': None, 'device_type': 'disk', 'encrypted': False, 'encryption_options': None, 'encryption_format': None, 'device_name': '/dev/sda', 'disk_bus': None, 'size': 0, 'image_id': '2e0c8bb4-0827-486c-98ce-7994555ffc23'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214206', 'volume_id': 'f319926d-a413-4adf-b108-e133bfe9bd26', 'name': 'volume-f319926d-a413-4adf-b108-e133bfe9bd26', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'a42e8c77-3cc3-4323-b1ee-2a95192910d5', 'attached_at': '', 'detached_at': '', 'volume_id': 'f319926d-a413-4adf-b108-e133bfe9bd26', 'serial': 'f319926d-a413-4adf-b108-e133bfe9bd26'}, 'mount_device': '/dev/sdb', 'guest_format': None, 'device_type': None, 'attachment_id': '2cf16f6b-b588-423d-8f11-2e0ac5647210', 'delete_on_termination': False, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=61857) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1053.386458] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Volume attach. Driver type: vmdk {{(pid=61857) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1053.386647] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214206', 'volume_id': 'f319926d-a413-4adf-b108-e133bfe9bd26', 'name': 'volume-f319926d-a413-4adf-b108-e133bfe9bd26', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'a42e8c77-3cc3-4323-b1ee-2a95192910d5', 'attached_at': '', 'detached_at': '', 'volume_id': 'f319926d-a413-4adf-b108-e133bfe9bd26', 'serial': 'f319926d-a413-4adf-b108-e133bfe9bd26'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1053.387459] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942fccfe-3c35-41f0-968c-b61348702f04 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.402603] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbafb04e-871c-446c-983d-1dcac8ffcdbb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.426406] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] volume-f319926d-a413-4adf-b108-e133bfe9bd26/volume-f319926d-a413-4adf-b108-e133bfe9bd26.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.426406] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-978614fa-c6a5-4039-98d8-20a0c1d335ae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.443921] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1053.443921] env[61857]: value = "task-951563" [ 1053.443921] env[61857]: _type = "Task" [ 1053.443921] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.454431] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951563, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.485714] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951561, 'name': Rename_Task, 'duration_secs': 0.149881} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.485994] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1053.486248] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74c6f1c8-0826-4aea-8d7b-1e8a7278b6a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.491873] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1053.491873] env[61857]: value = "task-951564" [ 1053.491873] env[61857]: _type = "Task" [ 1053.491873] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.502065] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951564, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.503469] env[61857]: DEBUG nova.compute.manager [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1053.678614] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951562, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063936} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.679049] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.679695] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0702db1-b5d3-4dc7-92ba-6eea4db0fbec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.701556] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 14df609d-678c-4713-8c9c-f80fabb61c79/14df609d-678c-4713-8c9c-f80fabb61c79.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.701862] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-116f7792-4c9f-402d-a6a1-694296cc7b01 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.721052] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1053.721052] env[61857]: value = "task-951565" [ 1053.721052] env[61857]: _type = "Task" [ 1053.721052] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.730366] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951565, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.953594] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951563, 'name': ReconfigVM_Task, 'duration_secs': 0.289861} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.953847] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Reconfigured VM instance instance-00000059 to attach disk [datastore1] volume-f319926d-a413-4adf-b108-e133bfe9bd26/volume-f319926d-a413-4adf-b108-e133bfe9bd26.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.958635] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d9066b2-02ef-43b2-8703-fd0531e3ad70 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.975186] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1053.975186] env[61857]: value = "task-951566" [ 1053.975186] env[61857]: _type = "Task" [ 1053.975186] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.988021] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951566, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.999519] env[61857]: DEBUG oslo_vmware.api [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951564, 'name': PowerOnVM_Task, 'duration_secs': 0.504441} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.999763] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1053.999973] env[61857]: INFO nova.compute.manager [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Took 10.52 seconds to spawn the instance on the hypervisor. [ 1054.000182] env[61857]: DEBUG nova.compute.manager [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1054.000971] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553874b1-0c2e-407e-818c-32e29b10e856 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.026630] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.026935] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.028447] env[61857]: INFO nova.compute.claims [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.231527] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951565, 'name': ReconfigVM_Task, 'duration_secs': 0.381933} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.231803] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 14df609d-678c-4713-8c9c-f80fabb61c79/14df609d-678c-4713-8c9c-f80fabb61c79.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.232444] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cc28498-8deb-4b3f-ae55-a4727bf28b4a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.239137] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1054.239137] env[61857]: value = "task-951567" [ 1054.239137] env[61857]: _type = "Task" [ 1054.239137] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.246596] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951567, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.484857] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951566, 'name': ReconfigVM_Task, 'duration_secs': 0.151497} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.485162] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214206', 'volume_id': 'f319926d-a413-4adf-b108-e133bfe9bd26', 'name': 'volume-f319926d-a413-4adf-b108-e133bfe9bd26', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'a42e8c77-3cc3-4323-b1ee-2a95192910d5', 'attached_at': '', 'detached_at': '', 'volume_id': 'f319926d-a413-4adf-b108-e133bfe9bd26', 'serial': 'f319926d-a413-4adf-b108-e133bfe9bd26'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1054.485752] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a455287d-fafd-4403-a08e-ce7a6d409736 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.492041] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1054.492041] env[61857]: value = "task-951568" [ 1054.492041] env[61857]: _type = "Task" [ 1054.492041] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.499787] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951568, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.521663] env[61857]: INFO nova.compute.manager [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Took 17.19 seconds to build instance. [ 1054.748482] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951567, 'name': Rename_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.002074] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951568, 'name': Rename_Task, 'duration_secs': 0.14544} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.002332] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1055.002604] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d5805dd-67fc-4d6d-8ccc-c0001bdcc6ac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.008105] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1055.008105] env[61857]: value = "task-951569" [ 1055.008105] env[61857]: _type = "Task" [ 1055.008105] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.015322] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.023904] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d6379e17-d4a7-4398-aab4-ac03bf262b6f tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.705s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.145042] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b3aa04-ddb6-4c2d-b6e1-4515f954436f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.154214] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25030c12-2710-4f79-a976-ccef8b0ddf32 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.185355] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0826921-f2be-4f7d-ab02-e89606b2227d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.192899] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16802cd7-079c-4ea7-b911-064a031269c9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.206187] env[61857]: DEBUG nova.compute.provider_tree [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.250524] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951567, 'name': Rename_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.518154] env[61857]: DEBUG oslo_vmware.api [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951569, 'name': PowerOnVM_Task, 'duration_secs': 0.466038} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.518396] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1055.617403] env[61857]: DEBUG nova.compute.manager [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1055.618437] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4924e780-0c0b-48d5-993b-1e0bab957524 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.710445] env[61857]: DEBUG nova.scheduler.client.report [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.749948] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951567, 'name': Rename_Task, 'duration_secs': 1.366492} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.751114] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1055.751114] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a9ac12c-ad7f-48d3-ac3a-bb94576df287 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.756388] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1055.756388] env[61857]: value = "task-951570" [ 1055.756388] env[61857]: _type = "Task" [ 1055.756388] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.764017] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.821195] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.821562] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.821768] env[61857]: DEBUG nova.compute.manager [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1055.822685] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ef29da-8fb7-44c7-b17b-03b3c6e9d667 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.830061] env[61857]: DEBUG nova.compute.manager [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61857) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1055.830925] env[61857]: DEBUG nova.objects.instance [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lazy-loading 'flavor' on Instance uuid 7ab9803b-8baa-4c49-b8df-61cd3b64b216 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.136832] env[61857]: DEBUG oslo_concurrency.lockutils [None req-41a717dc-1410-48da-862c-c0ba98758f6a tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 30.285s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.215447] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.188s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.216033] env[61857]: DEBUG nova.compute.manager [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1056.267473] env[61857]: DEBUG oslo_vmware.api [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951570, 'name': PowerOnVM_Task, 'duration_secs': 0.454855} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.267771] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1056.268059] env[61857]: INFO nova.compute.manager [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Took 10.56 seconds to spawn the instance on the hypervisor. [ 1056.268192] env[61857]: DEBUG nova.compute.manager [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.269063] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575e5cda-2fe7-40ba-b944-f57f4c3231e1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.337659] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1056.337976] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4fa436f-202d-48e5-a66b-4bd0e0cb9c12 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.346038] env[61857]: DEBUG oslo_vmware.api [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1056.346038] env[61857]: value = "task-951571" [ 1056.346038] env[61857]: _type = "Task" [ 1056.346038] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.357739] env[61857]: DEBUG oslo_vmware.api [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.721195] env[61857]: DEBUG nova.compute.utils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1056.723045] env[61857]: DEBUG nova.compute.manager [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1056.723045] env[61857]: DEBUG nova.network.neutron [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1056.764955] env[61857]: DEBUG nova.policy [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74f5499cbed541deaf1905605e10158f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62bd314daba846cd9ed3dc994dd1034b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1056.788746] env[61857]: INFO nova.compute.manager [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Took 18.89 seconds to build instance. [ 1056.856983] env[61857]: DEBUG oslo_vmware.api [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951571, 'name': PowerOffVM_Task, 'duration_secs': 0.208065} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.857585] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1056.857826] env[61857]: DEBUG nova.compute.manager [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.859339] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6eb3904-2b33-42ab-9ec8-97ae1ad74aa8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.024312] env[61857]: DEBUG nova.network.neutron [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Successfully created port: e1be624f-6a64-43aa-b429-788ac7363cc7 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1057.228914] env[61857]: DEBUG nova.compute.manager [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1057.291382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-269004f6-254e-4efa-a5c4-dc628ef5beed tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14df609d-678c-4713-8c9c-f80fabb61c79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.396s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.342525] env[61857]: DEBUG nova.compute.manager [req-8c9fb229-991b-4c7f-99b2-5fa114b25a1e req-7893eebe-0d78-41b8-8f88-4e50ff22db23 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received event network-changed-a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1057.342764] env[61857]: DEBUG nova.compute.manager [req-8c9fb229-991b-4c7f-99b2-5fa114b25a1e req-7893eebe-0d78-41b8-8f88-4e50ff22db23 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing instance network info cache due to event network-changed-a5cec797-c084-4a3f-828a-2624cd347348. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1057.342987] env[61857]: DEBUG oslo_concurrency.lockutils [req-8c9fb229-991b-4c7f-99b2-5fa114b25a1e req-7893eebe-0d78-41b8-8f88-4e50ff22db23 service nova] Acquiring lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.343485] env[61857]: DEBUG oslo_concurrency.lockutils [req-8c9fb229-991b-4c7f-99b2-5fa114b25a1e req-7893eebe-0d78-41b8-8f88-4e50ff22db23 service nova] Acquired lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.343670] env[61857]: DEBUG nova.network.neutron [req-8c9fb229-991b-4c7f-99b2-5fa114b25a1e req-7893eebe-0d78-41b8-8f88-4e50ff22db23 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing network info cache for port a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1057.370674] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c323bfec-fd61-4212-bba0-3b227b04db11 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.236837] env[61857]: DEBUG nova.compute.manager [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1058.239831] env[61857]: DEBUG nova.network.neutron [req-8c9fb229-991b-4c7f-99b2-5fa114b25a1e req-7893eebe-0d78-41b8-8f88-4e50ff22db23 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updated VIF entry in instance network info cache for port a5cec797-c084-4a3f-828a-2624cd347348. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1058.241621] env[61857]: DEBUG nova.network.neutron [req-8c9fb229-991b-4c7f-99b2-5fa114b25a1e req-7893eebe-0d78-41b8-8f88-4e50ff22db23 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updating instance_info_cache with network_info: [{"id": "a5cec797-c084-4a3f-828a-2624cd347348", "address": "fa:16:3e:e7:99:f4", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5cec797-c0", "ovs_interfaceid": "a5cec797-c084-4a3f-828a-2624cd347348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.273029] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1058.273319] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1058.273916] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.273916] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1058.273916] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.274147] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1058.274324] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1058.274495] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1058.274670] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1058.274874] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1058.275080] env[61857]: DEBUG nova.virt.hardware [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.276358] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1eb5734-054d-4cea-8dd8-a93a71e898b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.286248] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70a0171-246a-4457-b7f3-3bc7cb6f97fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.502193] env[61857]: DEBUG nova.compute.manager [req-aad78e91-e96b-4608-a729-296078853bbf req-b646c042-3870-4c60-a11d-4cf5d885d625 service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Received event network-vif-plugged-e1be624f-6a64-43aa-b429-788ac7363cc7 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1058.502436] env[61857]: DEBUG oslo_concurrency.lockutils [req-aad78e91-e96b-4608-a729-296078853bbf req-b646c042-3870-4c60-a11d-4cf5d885d625 service nova] Acquiring lock "01708c16-43f8-408a-ae6b-82a1ad34eafa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.502704] env[61857]: DEBUG oslo_concurrency.lockutils [req-aad78e91-e96b-4608-a729-296078853bbf req-b646c042-3870-4c60-a11d-4cf5d885d625 service nova] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.502854] env[61857]: DEBUG oslo_concurrency.lockutils [req-aad78e91-e96b-4608-a729-296078853bbf req-b646c042-3870-4c60-a11d-4cf5d885d625 service nova] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.503047] env[61857]: DEBUG nova.compute.manager [req-aad78e91-e96b-4608-a729-296078853bbf req-b646c042-3870-4c60-a11d-4cf5d885d625 service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] No waiting events found dispatching network-vif-plugged-e1be624f-6a64-43aa-b429-788ac7363cc7 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1058.503232] env[61857]: WARNING nova.compute.manager [req-aad78e91-e96b-4608-a729-296078853bbf req-b646c042-3870-4c60-a11d-4cf5d885d625 service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Received unexpected event network-vif-plugged-e1be624f-6a64-43aa-b429-788ac7363cc7 for instance with vm_state building and task_state spawning. [ 1058.591702] env[61857]: DEBUG nova.network.neutron [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Successfully updated port: e1be624f-6a64-43aa-b429-788ac7363cc7 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1058.707645] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.707928] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.708218] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.708484] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.708677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.711084] env[61857]: INFO nova.compute.manager [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Terminating instance [ 1058.713037] env[61857]: DEBUG nova.compute.manager [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1058.713284] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1058.714147] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0bbcde-6312-422e-88e2-0627d4d308c0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.722314] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1058.722553] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd67ddbb-6a35-47aa-8437-a0c95e31cdd9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.746454] env[61857]: DEBUG oslo_concurrency.lockutils [req-8c9fb229-991b-4c7f-99b2-5fa114b25a1e req-7893eebe-0d78-41b8-8f88-4e50ff22db23 service nova] Releasing lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.791927] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1058.792236] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1058.792488] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleting the datastore file [datastore2] 7ab9803b-8baa-4c49-b8df-61cd3b64b216 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1058.793034] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7fb23de9-c2af-4a9b-bf6d-abcfcb798f9b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.800098] env[61857]: DEBUG oslo_vmware.api [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1058.800098] env[61857]: value = "task-951573" [ 1058.800098] env[61857]: _type = "Task" [ 1058.800098] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.808323] env[61857]: DEBUG oslo_vmware.api [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951573, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.096321] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.096321] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.096321] env[61857]: DEBUG nova.network.neutron [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1059.310564] env[61857]: DEBUG oslo_vmware.api [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951573, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196384} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.310933] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1059.311044] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1059.311231] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1059.311448] env[61857]: INFO nova.compute.manager [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1059.311693] env[61857]: DEBUG oslo.service.loopingcall [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.311897] env[61857]: DEBUG nova.compute.manager [-] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1059.311998] env[61857]: DEBUG nova.network.neutron [-] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1059.377688] env[61857]: DEBUG nova.compute.manager [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Received event network-changed-a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1059.377936] env[61857]: DEBUG nova.compute.manager [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Refreshing instance network info cache due to event network-changed-a8571bda-024e-4d91-9417-6b5f64325f87. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1059.378227] env[61857]: DEBUG oslo_concurrency.lockutils [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] Acquiring lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.378393] env[61857]: DEBUG oslo_concurrency.lockutils [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] Acquired lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.378561] env[61857]: DEBUG nova.network.neutron [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Refreshing network info cache for port a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1059.626709] env[61857]: DEBUG nova.network.neutron [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1059.750109] env[61857]: DEBUG nova.network.neutron [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance_info_cache with network_info: [{"id": "e1be624f-6a64-43aa-b429-788ac7363cc7", "address": "fa:16:3e:67:11:cc", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1be624f-6a", "ovs_interfaceid": "e1be624f-6a64-43aa-b429-788ac7363cc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.079483] env[61857]: DEBUG nova.network.neutron [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updated VIF entry in instance network info cache for port a8571bda-024e-4d91-9417-6b5f64325f87. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1060.079917] env[61857]: DEBUG nova.network.neutron [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updating instance_info_cache with network_info: [{"id": "a8571bda-024e-4d91-9417-6b5f64325f87", "address": "fa:16:3e:e8:b5:c8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8571bda-02", "ovs_interfaceid": "a8571bda-024e-4d91-9417-6b5f64325f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.094398] env[61857]: DEBUG nova.network.neutron [-] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.252880] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.253222] env[61857]: DEBUG nova.compute.manager [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Instance network_info: |[{"id": "e1be624f-6a64-43aa-b429-788ac7363cc7", "address": "fa:16:3e:67:11:cc", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1be624f-6a", "ovs_interfaceid": "e1be624f-6a64-43aa-b429-788ac7363cc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1060.253819] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:11:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f678cd81-6d15-43d5-aab7-d7eedc2ef2d5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e1be624f-6a64-43aa-b429-788ac7363cc7', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1060.261149] env[61857]: DEBUG oslo.service.loopingcall [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.261382] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1060.261618] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b4c926c-e286-4f25-a46c-eb871e9f5b44 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.280945] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1060.280945] env[61857]: value = "task-951574" [ 1060.280945] env[61857]: _type = "Task" [ 1060.280945] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.288554] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951574, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.537823] env[61857]: DEBUG nova.compute.manager [req-f9529591-744b-445f-abb7-c0486e68f07d req-ace6ef7b-c0ab-4351-9a40-4e5c94c45ef7 service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Received event network-changed-e1be624f-6a64-43aa-b429-788ac7363cc7 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1060.538044] env[61857]: DEBUG nova.compute.manager [req-f9529591-744b-445f-abb7-c0486e68f07d req-ace6ef7b-c0ab-4351-9a40-4e5c94c45ef7 service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Refreshing instance network info cache due to event network-changed-e1be624f-6a64-43aa-b429-788ac7363cc7. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1060.538943] env[61857]: DEBUG oslo_concurrency.lockutils [req-f9529591-744b-445f-abb7-c0486e68f07d req-ace6ef7b-c0ab-4351-9a40-4e5c94c45ef7 service nova] Acquiring lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.538943] env[61857]: DEBUG oslo_concurrency.lockutils [req-f9529591-744b-445f-abb7-c0486e68f07d req-ace6ef7b-c0ab-4351-9a40-4e5c94c45ef7 service nova] Acquired lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.538943] env[61857]: DEBUG nova.network.neutron [req-f9529591-744b-445f-abb7-c0486e68f07d req-ace6ef7b-c0ab-4351-9a40-4e5c94c45ef7 service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Refreshing network info cache for port e1be624f-6a64-43aa-b429-788ac7363cc7 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1060.582336] env[61857]: DEBUG oslo_concurrency.lockutils [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] Releasing lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.582707] env[61857]: DEBUG nova.compute.manager [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Received event network-changed-a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1060.582933] env[61857]: DEBUG nova.compute.manager [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Refreshing instance network info cache due to event network-changed-a8571bda-024e-4d91-9417-6b5f64325f87. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1060.583182] env[61857]: DEBUG oslo_concurrency.lockutils [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] Acquiring lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.583338] env[61857]: DEBUG oslo_concurrency.lockutils [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] Acquired lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.583507] env[61857]: DEBUG nova.network.neutron [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Refreshing network info cache for port a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1060.597016] env[61857]: INFO nova.compute.manager [-] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Took 1.28 seconds to deallocate network for instance. [ 1060.791292] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951574, 'name': CreateVM_Task, 'duration_secs': 0.360265} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.791509] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1060.792243] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.792423] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.792802] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1060.793076] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3fd43c7-e1b6-45e0-b1c2-406a2b61815b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.797348] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1060.797348] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52f76b5a-d220-b5cd-bc50-f8d0404479b1" [ 1060.797348] env[61857]: _type = "Task" [ 1060.797348] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.805233] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f76b5a-d220-b5cd-bc50-f8d0404479b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.104013] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.104317] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.104542] env[61857]: DEBUG nova.objects.instance [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lazy-loading 'resources' on Instance uuid 7ab9803b-8baa-4c49-b8df-61cd3b64b216 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.286447] env[61857]: DEBUG nova.network.neutron [req-f9529591-744b-445f-abb7-c0486e68f07d req-ace6ef7b-c0ab-4351-9a40-4e5c94c45ef7 service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updated VIF entry in instance network info cache for port e1be624f-6a64-43aa-b429-788ac7363cc7. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1061.286827] env[61857]: DEBUG nova.network.neutron [req-f9529591-744b-445f-abb7-c0486e68f07d req-ace6ef7b-c0ab-4351-9a40-4e5c94c45ef7 service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance_info_cache with network_info: [{"id": "e1be624f-6a64-43aa-b429-788ac7363cc7", "address": "fa:16:3e:67:11:cc", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1be624f-6a", "ovs_interfaceid": "e1be624f-6a64-43aa-b429-788ac7363cc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.295451] env[61857]: DEBUG nova.network.neutron [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updated VIF entry in instance network info cache for port a8571bda-024e-4d91-9417-6b5f64325f87. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1061.295789] env[61857]: DEBUG nova.network.neutron [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updating instance_info_cache with network_info: [{"id": "a8571bda-024e-4d91-9417-6b5f64325f87", "address": "fa:16:3e:e8:b5:c8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8571bda-02", "ovs_interfaceid": "a8571bda-024e-4d91-9417-6b5f64325f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.308231] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f76b5a-d220-b5cd-bc50-f8d0404479b1, 'name': SearchDatastore_Task, 'duration_secs': 0.008338} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.308555] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.308789] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.309034] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.309189] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.309404] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.309669] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9aa5e62e-b398-44a4-bd3f-9f3940cbc0a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.317389] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1061.317570] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1061.318278] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7597b0e2-4563-4f0c-96ee-4d24273b1e56 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.323334] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1061.323334] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fd8c8c-6cfa-8f54-33d7-d881ac93285f" [ 1061.323334] env[61857]: _type = "Task" [ 1061.323334] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.330720] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fd8c8c-6cfa-8f54-33d7-d881ac93285f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.704651] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6357db-597f-44f6-b456-8865e5191a38 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.712164] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4b8a0b-afd7-4920-a966-40b912d8bd0a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.742484] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-690b6219-cf4f-415b-a324-18caed30cb23 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.749204] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398061fc-2e0e-4fdc-b794-822d8cdaba43 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.761838] env[61857]: DEBUG nova.compute.provider_tree [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.789469] env[61857]: DEBUG oslo_concurrency.lockutils [req-f9529591-744b-445f-abb7-c0486e68f07d req-ace6ef7b-c0ab-4351-9a40-4e5c94c45ef7 service nova] Releasing lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.789701] env[61857]: DEBUG nova.compute.manager [req-f9529591-744b-445f-abb7-c0486e68f07d req-ace6ef7b-c0ab-4351-9a40-4e5c94c45ef7 service nova] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Received event network-vif-deleted-63f7bcf4-8432-4c8a-beb5-a3b7c1b25d31 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1061.798323] env[61857]: DEBUG oslo_concurrency.lockutils [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] Releasing lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.798534] env[61857]: DEBUG nova.compute.manager [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received event network-changed-a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1061.798706] env[61857]: DEBUG nova.compute.manager [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing instance network info cache due to event network-changed-a5cec797-c084-4a3f-828a-2624cd347348. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1061.798912] env[61857]: DEBUG oslo_concurrency.lockutils [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] Acquiring lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.799075] env[61857]: DEBUG oslo_concurrency.lockutils [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] Acquired lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.799247] env[61857]: DEBUG nova.network.neutron [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing network info cache for port a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1061.832754] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fd8c8c-6cfa-8f54-33d7-d881ac93285f, 'name': SearchDatastore_Task, 'duration_secs': 0.008678} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.833473] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e1f7ae1-14a8-4c01-9067-8f57bf79b0de {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.838027] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1061.838027] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5211ba48-625d-8747-9ea9-8a6cb8bb1f12" [ 1061.838027] env[61857]: _type = "Task" [ 1061.838027] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.845706] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5211ba48-625d-8747-9ea9-8a6cb8bb1f12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.264749] env[61857]: DEBUG nova.scheduler.client.report [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1062.347905] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5211ba48-625d-8747-9ea9-8a6cb8bb1f12, 'name': SearchDatastore_Task, 'duration_secs': 0.00944} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.348108] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.348375] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 01708c16-43f8-408a-ae6b-82a1ad34eafa/01708c16-43f8-408a-ae6b-82a1ad34eafa.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1062.348922] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b417c48a-eea9-476a-843e-7161daac9246 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.355336] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1062.355336] env[61857]: value = "task-951575" [ 1062.355336] env[61857]: _type = "Task" [ 1062.355336] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.362994] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951575, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.513913] env[61857]: DEBUG nova.network.neutron [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updated VIF entry in instance network info cache for port a5cec797-c084-4a3f-828a-2624cd347348. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1062.514443] env[61857]: DEBUG nova.network.neutron [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updating instance_info_cache with network_info: [{"id": "a5cec797-c084-4a3f-828a-2624cd347348", "address": "fa:16:3e:e7:99:f4", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5cec797-c0", "ovs_interfaceid": "a5cec797-c084-4a3f-828a-2624cd347348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.770633] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.790699] env[61857]: INFO nova.scheduler.client.report [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted allocations for instance 7ab9803b-8baa-4c49-b8df-61cd3b64b216 [ 1062.867539] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951575, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.410987} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.867851] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] 01708c16-43f8-408a-ae6b-82a1ad34eafa/01708c16-43f8-408a-ae6b-82a1ad34eafa.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1062.868110] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1062.868374] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d3a2106-6d17-4fed-98f7-cde7d8237a10 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.875306] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1062.875306] env[61857]: value = "task-951576" [ 1062.875306] env[61857]: _type = "Task" [ 1062.875306] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.886686] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951576, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.017838] env[61857]: DEBUG oslo_concurrency.lockutils [req-43143e7e-fb7c-48b3-ba88-e53bbf8f437c req-08327d21-6406-47c5-acbe-48307f58478f service nova] Releasing lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.300361] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8b0af5e8-7d94-4e7f-92f6-a8a806e177dd tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "7ab9803b-8baa-4c49-b8df-61cd3b64b216" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.592s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.385767] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951576, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063522} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.386074] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1063.386862] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23b70b4-72e2-4e5c-95bf-fb0691bc3ecb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.409396] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 01708c16-43f8-408a-ae6b-82a1ad34eafa/01708c16-43f8-408a-ae6b-82a1ad34eafa.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.409727] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3db29581-24f6-465b-9b82-60b861acd5f6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.429554] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1063.429554] env[61857]: value = "task-951577" [ 1063.429554] env[61857]: _type = "Task" [ 1063.429554] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.437738] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951577, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.940104] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951577, 'name': ReconfigVM_Task, 'duration_secs': 0.27031} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.940480] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 01708c16-43f8-408a-ae6b-82a1ad34eafa/01708c16-43f8-408a-ae6b-82a1ad34eafa.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.941138] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5ac3867-cb47-483a-9e16-bb1344393c1b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.947594] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1063.947594] env[61857]: value = "task-951578" [ 1063.947594] env[61857]: _type = "Task" [ 1063.947594] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.956255] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951578, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.201505] env[61857]: DEBUG nova.compute.manager [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1064.202468] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be6af73-b94c-4bfb-ac19-30ab70f82a4c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.457291] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951578, 'name': Rename_Task, 'duration_secs': 0.153877} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.457634] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1064.457889] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e4e8a5a-38a9-49e7-aea6-da1a5906010b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.464315] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1064.464315] env[61857]: value = "task-951579" [ 1064.464315] env[61857]: _type = "Task" [ 1064.464315] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.473106] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951579, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.611403] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.611686] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.611900] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.612105] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.612287] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.614659] env[61857]: INFO nova.compute.manager [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Terminating instance [ 1064.616572] env[61857]: DEBUG nova.compute.manager [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1064.616785] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1064.617606] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224bbcb7-3e31-484f-a75d-1ad632299ddb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.625663] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1064.625902] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-611c5e71-1dff-4b05-b9a8-a0882d8dc94e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.632769] env[61857]: DEBUG oslo_vmware.api [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1064.632769] env[61857]: value = "task-951580" [ 1064.632769] env[61857]: _type = "Task" [ 1064.632769] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.640225] env[61857]: DEBUG oslo_vmware.api [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951580, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.715865] env[61857]: INFO nova.compute.manager [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] instance snapshotting [ 1064.716760] env[61857]: DEBUG nova.objects.instance [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'flavor' on Instance uuid 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.974672] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951579, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.143175] env[61857]: DEBUG oslo_vmware.api [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951580, 'name': PowerOffVM_Task, 'duration_secs': 0.175331} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.143387] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1065.143558] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1065.143817] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e03cb79-b7ba-4018-b554-ed29c445e42d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.207339] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1065.207486] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1065.207785] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleting the datastore file [datastore1] 18d9307d-6a20-47b4-b4e9-176f9a8c33cc {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1065.207941] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f98e36fd-1009-4fd1-b69a-39dc1cadca5b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.214053] env[61857]: DEBUG oslo_vmware.api [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for the task: (returnval){ [ 1065.214053] env[61857]: value = "task-951582" [ 1065.214053] env[61857]: _type = "Task" [ 1065.214053] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.223245] env[61857]: DEBUG oslo_vmware.api [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951582, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.224212] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b144ada-e705-4873-be53-aa8500d13408 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.241776] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f4ca8f-a6b5-43bc-9ab6-8b0960b8f6f7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.474965] env[61857]: DEBUG oslo_vmware.api [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951579, 'name': PowerOnVM_Task, 'duration_secs': 0.693356} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.475202] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1065.475411] env[61857]: INFO nova.compute.manager [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Took 7.24 seconds to spawn the instance on the hypervisor. [ 1065.475593] env[61857]: DEBUG nova.compute.manager [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.476350] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9d3044-fb84-4f77-b9c8-6277fca235a5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.725019] env[61857]: DEBUG oslo_vmware.api [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Task: {'id': task-951582, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179572} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.725287] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1065.725489] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1065.725684] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1065.725895] env[61857]: INFO nova.compute.manager [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1065.726239] env[61857]: DEBUG oslo.service.loopingcall [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1065.726546] env[61857]: DEBUG nova.compute.manager [-] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1065.726694] env[61857]: DEBUG nova.network.neutron [-] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1065.752773] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1065.753103] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-840e5552-4ebd-4d71-8fde-e36fff256619 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.761929] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1065.761929] env[61857]: value = "task-951583" [ 1065.761929] env[61857]: _type = "Task" [ 1065.761929] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.771717] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951583, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.992045] env[61857]: INFO nova.compute.manager [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Took 11.98 seconds to build instance. [ 1065.995401] env[61857]: DEBUG nova.compute.manager [req-9cb3dc29-9cbf-4c1a-ac6f-be22c07c7b50 req-9198c954-a331-42d4-8c15-34280696063c service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Received event network-vif-deleted-527bd220-3cfd-4df1-b3c6-014e4bd05cc9 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1065.995605] env[61857]: INFO nova.compute.manager [req-9cb3dc29-9cbf-4c1a-ac6f-be22c07c7b50 req-9198c954-a331-42d4-8c15-34280696063c service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Neutron deleted interface 527bd220-3cfd-4df1-b3c6-014e4bd05cc9; detaching it from the instance and deleting it from the info cache [ 1065.995784] env[61857]: DEBUG nova.network.neutron [req-9cb3dc29-9cbf-4c1a-ac6f-be22c07c7b50 req-9198c954-a331-42d4-8c15-34280696063c service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.272079] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951583, 'name': CreateSnapshot_Task, 'duration_secs': 0.445665} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.272385] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1066.273162] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a06ec6-d684-4127-a5c0-914720e332e5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.468339] env[61857]: DEBUG nova.network.neutron [-] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.498760] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fcedd649-c2d8-4656-b954-834469d8e2eb tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.498s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.499024] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab7f2525-01fe-426e-9770-52d13e57e8bd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.510395] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9835fbb-cccc-4891-970f-5b5ebf03214e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.536764] env[61857]: DEBUG nova.compute.manager [req-9cb3dc29-9cbf-4c1a-ac6f-be22c07c7b50 req-9198c954-a331-42d4-8c15-34280696063c service nova] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Detach interface failed, port_id=527bd220-3cfd-4df1-b3c6-014e4bd05cc9, reason: Instance 18d9307d-6a20-47b4-b4e9-176f9a8c33cc could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1066.790845] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1066.791132] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5ade8e11-f14c-44e2-bc45-cfed327ff694 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.799479] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1066.799479] env[61857]: value = "task-951584" [ 1066.799479] env[61857]: _type = "Task" [ 1066.799479] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.807168] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951584, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.972612] env[61857]: INFO nova.compute.manager [-] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Took 1.25 seconds to deallocate network for instance. [ 1067.075278] env[61857]: DEBUG nova.compute.manager [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Stashing vm_state: active {{(pid=61857) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1067.310225] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951584, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.482296] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.482581] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.482803] env[61857]: DEBUG nova.objects.instance [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lazy-loading 'resources' on Instance uuid 18d9307d-6a20-47b4-b4e9-176f9a8c33cc {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.598264] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.810175] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951584, 'name': CloneVM_Task} progress is 95%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.099687] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67eedc8b-1372-4485-ad4a-df7e5b1eabbf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.107630] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5083b8c1-c835-45ca-9829-d7a2decdaa05 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.140206] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8cc34b-e311-4bcb-8343-ea3bb1d1f9b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.150231] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41ba197-94ed-4929-8848-1d75d8ceca80 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.164074] env[61857]: DEBUG nova.compute.provider_tree [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.221260] env[61857]: DEBUG nova.compute.manager [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Stashing vm_state: active {{(pid=61857) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1068.310952] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951584, 'name': CloneVM_Task, 'duration_secs': 1.116914} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.311266] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Created linked-clone VM from snapshot [ 1068.312058] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba550e5-cc84-4a21-906d-6f21301ffa17 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.320083] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Uploading image 3448d3d7-419d-4115-8a4c-1b26372f70aa {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1068.341329] env[61857]: DEBUG oslo_vmware.rw_handles [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1068.341329] env[61857]: value = "vm-214218" [ 1068.341329] env[61857]: _type = "VirtualMachine" [ 1068.341329] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1068.341669] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0e25d262-e252-4af8-8bdc-8dab240b8a97 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.348241] env[61857]: DEBUG oslo_vmware.rw_handles [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lease: (returnval){ [ 1068.348241] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b1015e-5af2-ee98-099c-56558eb7ce02" [ 1068.348241] env[61857]: _type = "HttpNfcLease" [ 1068.348241] env[61857]: } obtained for exporting VM: (result){ [ 1068.348241] env[61857]: value = "vm-214218" [ 1068.348241] env[61857]: _type = "VirtualMachine" [ 1068.348241] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1068.348509] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the lease: (returnval){ [ 1068.348509] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b1015e-5af2-ee98-099c-56558eb7ce02" [ 1068.348509] env[61857]: _type = "HttpNfcLease" [ 1068.348509] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1068.354820] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1068.354820] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b1015e-5af2-ee98-099c-56558eb7ce02" [ 1068.354820] env[61857]: _type = "HttpNfcLease" [ 1068.354820] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1068.640275] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "interface-14844876-96ab-481d-980d-4133e89d7d4d-5b14caa2-9c46-4c97-a303-7157978d0b97" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.640606] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-14844876-96ab-481d-980d-4133e89d7d4d-5b14caa2-9c46-4c97-a303-7157978d0b97" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.641392] env[61857]: DEBUG nova.objects.instance [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'flavor' on Instance uuid 14844876-96ab-481d-980d-4133e89d7d4d {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.667408] env[61857]: DEBUG nova.scheduler.client.report [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.737650] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.856018] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1068.856018] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b1015e-5af2-ee98-099c-56558eb7ce02" [ 1068.856018] env[61857]: _type = "HttpNfcLease" [ 1068.856018] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1068.856329] env[61857]: DEBUG oslo_vmware.rw_handles [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1068.856329] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b1015e-5af2-ee98-099c-56558eb7ce02" [ 1068.856329] env[61857]: _type = "HttpNfcLease" [ 1068.856329] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1068.857049] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3c13bb-b656-4939-b8a2-13c52fd1cbfc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.864135] env[61857]: DEBUG oslo_vmware.rw_handles [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a6eac6-3519-4c0b-880d-34dcc0350bcb/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1068.864313] env[61857]: DEBUG oslo_vmware.rw_handles [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a6eac6-3519-4c0b-880d-34dcc0350bcb/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1068.951214] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6fc6e256-4f08-4ac7-b754-9aaa9579305d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.172260] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.690s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.175032] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.576s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.194130] env[61857]: INFO nova.scheduler.client.report [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Deleted allocations for instance 18d9307d-6a20-47b4-b4e9-176f9a8c33cc [ 1069.250205] env[61857]: DEBUG nova.objects.instance [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'pci_requests' on Instance uuid 14844876-96ab-481d-980d-4133e89d7d4d {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.679725] env[61857]: INFO nova.compute.claims [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.703964] env[61857]: DEBUG oslo_concurrency.lockutils [None req-fe7b4efa-073b-4f4a-b56c-e43ec662bd48 tempest-ServersTestJSON-1566382171 tempest-ServersTestJSON-1566382171-project-member] Lock "18d9307d-6a20-47b4-b4e9-176f9a8c33cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.092s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.753284] env[61857]: DEBUG nova.objects.base [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Object Instance<14844876-96ab-481d-980d-4133e89d7d4d> lazy-loaded attributes: flavor,pci_requests {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1069.753690] env[61857]: DEBUG nova.network.neutron [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1069.862481] env[61857]: DEBUG nova.policy [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9056d344ff614950a60983450a2a40ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5653ed6ab0a948ef89bb03cf7426dd10', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1070.188608] env[61857]: INFO nova.compute.resource_tracker [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating resource usage from migration b12e16cf-7bb8-4a79-bed3-ad040c5e8f5f [ 1070.316016] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6482fd1d-92c5-4367-b672-1a440a707d37 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.323449] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608210ee-3d27-489d-b490-a9d86b7e7e5b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.354018] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c666620c-cd8c-44d6-a9e4-00ea6ed56f21 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.362175] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008c7f4a-dee4-4efd-9828-6202b8bb8d1b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.377459] env[61857]: DEBUG nova.compute.provider_tree [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.881335] env[61857]: DEBUG nova.scheduler.client.report [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.386127] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.212s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.386381] env[61857]: INFO nova.compute.manager [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Migrating [ 1071.399101] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.661s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.468749] env[61857]: DEBUG nova.compute.manager [req-6003b69c-daea-411e-bee2-b4064010c183 req-b5ed1fb6-eb9a-44ab-8ca6-7e2cf261cfcb service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received event network-vif-plugged-5b14caa2-9c46-4c97-a303-7157978d0b97 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1071.469172] env[61857]: DEBUG oslo_concurrency.lockutils [req-6003b69c-daea-411e-bee2-b4064010c183 req-b5ed1fb6-eb9a-44ab-8ca6-7e2cf261cfcb service nova] Acquiring lock "14844876-96ab-481d-980d-4133e89d7d4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.469447] env[61857]: DEBUG oslo_concurrency.lockutils [req-6003b69c-daea-411e-bee2-b4064010c183 req-b5ed1fb6-eb9a-44ab-8ca6-7e2cf261cfcb service nova] Lock "14844876-96ab-481d-980d-4133e89d7d4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.471124] env[61857]: DEBUG oslo_concurrency.lockutils [req-6003b69c-daea-411e-bee2-b4064010c183 req-b5ed1fb6-eb9a-44ab-8ca6-7e2cf261cfcb service nova] Lock "14844876-96ab-481d-980d-4133e89d7d4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.471124] env[61857]: DEBUG nova.compute.manager [req-6003b69c-daea-411e-bee2-b4064010c183 req-b5ed1fb6-eb9a-44ab-8ca6-7e2cf261cfcb service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] No waiting events found dispatching network-vif-plugged-5b14caa2-9c46-4c97-a303-7157978d0b97 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1071.471124] env[61857]: WARNING nova.compute.manager [req-6003b69c-daea-411e-bee2-b4064010c183 req-b5ed1fb6-eb9a-44ab-8ca6-7e2cf261cfcb service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received unexpected event network-vif-plugged-5b14caa2-9c46-4c97-a303-7157978d0b97 for instance with vm_state active and task_state None. [ 1071.560739] env[61857]: DEBUG nova.network.neutron [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Successfully updated port: 5b14caa2-9c46-4c97-a303-7157978d0b97 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1071.908732] env[61857]: INFO nova.compute.claims [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1071.912388] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.912582] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.912739] env[61857]: DEBUG nova.network.neutron [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1072.063582] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.063830] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.063999] env[61857]: DEBUG nova.network.neutron [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1072.419710] env[61857]: INFO nova.compute.resource_tracker [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating resource usage from migration 75be059d-4418-4cfb-8c2d-feb19efd97a6 [ 1072.539245] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79337a06-9272-45dd-84dc-300c07784c6a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.549311] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5018d745-564d-4389-b506-5190bef52386 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.583954] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7501282b-01b8-465b-b7e8-02f02727772a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.592508] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7f47fc-a77b-4a5e-b8ff-03e635ca0f93 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.607508] env[61857]: DEBUG nova.compute.provider_tree [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.633178] env[61857]: WARNING nova.network.neutron [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] 30b7d27c-744c-46a6-8769-b1b1c95306e4 already exists in list: networks containing: ['30b7d27c-744c-46a6-8769-b1b1c95306e4']. ignoring it [ 1072.659419] env[61857]: DEBUG nova.network.neutron [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance_info_cache with network_info: [{"id": "e1be624f-6a64-43aa-b429-788ac7363cc7", "address": "fa:16:3e:67:11:cc", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1be624f-6a", "ovs_interfaceid": "e1be624f-6a64-43aa-b429-788ac7363cc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.937613] env[61857]: DEBUG nova.network.neutron [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updating instance_info_cache with network_info: [{"id": "a5cec797-c084-4a3f-828a-2624cd347348", "address": "fa:16:3e:e7:99:f4", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5cec797-c0", "ovs_interfaceid": "a5cec797-c084-4a3f-828a-2624cd347348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5b14caa2-9c46-4c97-a303-7157978d0b97", "address": "fa:16:3e:17:98:ef", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b14caa2-9c", "ovs_interfaceid": "5b14caa2-9c46-4c97-a303-7157978d0b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.111116] env[61857]: DEBUG nova.scheduler.client.report [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1073.163422] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.440825] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.441559] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.441759] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.442726] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ad6d2a-686c-48b0-a011-ee1f173425c4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.459958] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1073.460211] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1073.460404] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1073.460612] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1073.460770] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1073.460970] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1073.461206] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1073.461390] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1073.461548] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1073.461716] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1073.461895] env[61857]: DEBUG nova.virt.hardware [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.468043] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Reconfiguring VM to attach interface {{(pid=61857) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1073.468380] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51d5c239-92e9-483e-bb39-c22756cee5d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.485584] env[61857]: DEBUG oslo_vmware.api [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1073.485584] env[61857]: value = "task-951586" [ 1073.485584] env[61857]: _type = "Task" [ 1073.485584] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.493532] env[61857]: DEBUG oslo_vmware.api [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951586, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.496394] env[61857]: DEBUG nova.compute.manager [req-581aca4b-fe56-42e2-8b0c-2a979428e923 req-5efe36b9-9c8b-487b-94f6-e5d7c567a354 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received event network-changed-5b14caa2-9c46-4c97-a303-7157978d0b97 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1073.496543] env[61857]: DEBUG nova.compute.manager [req-581aca4b-fe56-42e2-8b0c-2a979428e923 req-5efe36b9-9c8b-487b-94f6-e5d7c567a354 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing instance network info cache due to event network-changed-5b14caa2-9c46-4c97-a303-7157978d0b97. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1073.496783] env[61857]: DEBUG oslo_concurrency.lockutils [req-581aca4b-fe56-42e2-8b0c-2a979428e923 req-5efe36b9-9c8b-487b-94f6-e5d7c567a354 service nova] Acquiring lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.496888] env[61857]: DEBUG oslo_concurrency.lockutils [req-581aca4b-fe56-42e2-8b0c-2a979428e923 req-5efe36b9-9c8b-487b-94f6-e5d7c567a354 service nova] Acquired lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.497077] env[61857]: DEBUG nova.network.neutron [req-581aca4b-fe56-42e2-8b0c-2a979428e923 req-5efe36b9-9c8b-487b-94f6-e5d7c567a354 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing network info cache for port 5b14caa2-9c46-4c97-a303-7157978d0b97 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1073.617091] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.218s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.617343] env[61857]: INFO nova.compute.manager [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Migrating [ 1073.996243] env[61857]: DEBUG oslo_vmware.api [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951586, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.134423] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.134610] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.134858] env[61857]: DEBUG nova.network.neutron [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1074.217410] env[61857]: DEBUG nova.network.neutron [req-581aca4b-fe56-42e2-8b0c-2a979428e923 req-5efe36b9-9c8b-487b-94f6-e5d7c567a354 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updated VIF entry in instance network info cache for port 5b14caa2-9c46-4c97-a303-7157978d0b97. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1074.217907] env[61857]: DEBUG nova.network.neutron [req-581aca4b-fe56-42e2-8b0c-2a979428e923 req-5efe36b9-9c8b-487b-94f6-e5d7c567a354 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updating instance_info_cache with network_info: [{"id": "a5cec797-c084-4a3f-828a-2624cd347348", "address": "fa:16:3e:e7:99:f4", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5cec797-c0", "ovs_interfaceid": "a5cec797-c084-4a3f-828a-2624cd347348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5b14caa2-9c46-4c97-a303-7157978d0b97", "address": "fa:16:3e:17:98:ef", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b14caa2-9c", "ovs_interfaceid": "5b14caa2-9c46-4c97-a303-7157978d0b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.497241] env[61857]: DEBUG oslo_vmware.api [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951586, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.679710] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80ab224-78e4-46c9-b8a6-7837d3d96b1a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.699595] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance '01708c16-43f8-408a-ae6b-82a1ad34eafa' progress to 0 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1074.721872] env[61857]: DEBUG oslo_concurrency.lockutils [req-581aca4b-fe56-42e2-8b0c-2a979428e923 req-5efe36b9-9c8b-487b-94f6-e5d7c567a354 service nova] Releasing lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.860565] env[61857]: DEBUG nova.network.neutron [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance_info_cache with network_info: [{"id": "83ff6e16-8b1b-4391-9ef1-57600502688c", "address": "fa:16:3e:46:c4:d7", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83ff6e16-8b", "ovs_interfaceid": "83ff6e16-8b1b-4391-9ef1-57600502688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.997280] env[61857]: DEBUG oslo_vmware.api [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951586, 'name': ReconfigVM_Task, 'duration_secs': 1.085141} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.997823] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.998061] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Reconfigured VM to attach interface {{(pid=61857) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1075.206474] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1075.206827] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57e6adcc-aab0-4d90-b129-8d4fc05fff59 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.214763] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1075.214763] env[61857]: value = "task-951587" [ 1075.214763] env[61857]: _type = "Task" [ 1075.214763] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.222614] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.363323] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.502643] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8aed0551-9794-4f37-bec7-6ac6b9bc0416 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-14844876-96ab-481d-980d-4133e89d7d4d-5b14caa2-9c46-4c97-a303-7157978d0b97" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.862s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.723263] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951587, 'name': PowerOffVM_Task, 'duration_secs': 0.234079} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.723620] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1075.723822] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance '01708c16-43f8-408a-ae6b-82a1ad34eafa' progress to 17 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1076.230613] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1076.231032] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1076.231114] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1076.231344] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1076.231561] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1076.231738] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1076.231975] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1076.232217] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1076.232415] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1076.232626] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1076.232834] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1076.238824] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03c05261-9831-4d6b-bb20-2ac35bf2a435 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.257286] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1076.257286] env[61857]: value = "task-951588" [ 1076.257286] env[61857]: _type = "Task" [ 1076.257286] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.267798] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951588, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.767348] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951588, 'name': ReconfigVM_Task, 'duration_secs': 0.211544} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.767698] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance '01708c16-43f8-408a-ae6b-82a1ad34eafa' progress to 33 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1076.798112] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "interface-14844876-96ab-481d-980d-4133e89d7d4d-5b14caa2-9c46-4c97-a303-7157978d0b97" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.798447] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-14844876-96ab-481d-980d-4133e89d7d4d-5b14caa2-9c46-4c97-a303-7157978d0b97" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.880600] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2389fd1-99f8-49e0-aac4-b1076b9eb641 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.899104] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance '04438fc0-b5a1-4e29-84d5-4d84d332c8e4' progress to 0 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1077.274891] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.275167] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.275334] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.275526] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.275679] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.275831] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.276053] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.276242] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.276451] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.276623] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.276814] env[61857]: DEBUG nova.virt.hardware [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.282891] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1077.283213] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ebe9774-f1f9-4b10-a127-1b4751939cd2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.302813] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.302937] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.303272] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1077.303272] env[61857]: value = "task-951589" [ 1077.303272] env[61857]: _type = "Task" [ 1077.303272] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.303996] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e899ed4-fcf6-416f-be2f-41bba0b1ebaf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.315371] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951589, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.329609] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606639ec-1c23-4805-b148-ef859b9dc167 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.357576] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Reconfiguring VM to detach interface {{(pid=61857) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1077.358249] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ead42ba-018e-4aa7-ad59-0daf65950335 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.378582] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1077.378582] env[61857]: value = "task-951590" [ 1077.378582] env[61857]: _type = "Task" [ 1077.378582] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.386675] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.394537] env[61857]: DEBUG oslo_vmware.rw_handles [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a6eac6-3519-4c0b-880d-34dcc0350bcb/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1077.395478] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a3a835-44e6-4def-879c-c72682dbd57b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.401638] env[61857]: DEBUG oslo_vmware.rw_handles [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a6eac6-3519-4c0b-880d-34dcc0350bcb/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1077.401886] env[61857]: ERROR oslo_vmware.rw_handles [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a6eac6-3519-4c0b-880d-34dcc0350bcb/disk-0.vmdk due to incomplete transfer. [ 1077.402256] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-614d8361-f0f5-47f9-9d5d-f75072555c2c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.406092] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1077.406368] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d0781cc-dc88-4a5a-a231-44d7064544d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.413483] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1077.413483] env[61857]: value = "task-951591" [ 1077.413483] env[61857]: _type = "Task" [ 1077.413483] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.414645] env[61857]: DEBUG oslo_vmware.rw_handles [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a6eac6-3519-4c0b-880d-34dcc0350bcb/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1077.414840] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Uploaded image 3448d3d7-419d-4115-8a4c-1b26372f70aa to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1077.417215] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1077.420399] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a570f6ac-8191-40d6-8805-a9903477442c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.426315] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951591, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.427745] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1077.427745] env[61857]: value = "task-951592" [ 1077.427745] env[61857]: _type = "Task" [ 1077.427745] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.438408] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951592, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.817217] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951589, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.888279] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.924128] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951591, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.935642] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951592, 'name': Destroy_Task, 'duration_secs': 0.337497} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.935903] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Destroyed the VM [ 1077.936161] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1077.936400] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1938300a-678e-4c3c-a0c6-cf5a521e6db3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.942466] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1077.942466] env[61857]: value = "task-951593" [ 1077.942466] env[61857]: _type = "Task" [ 1077.942466] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.949626] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951593, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.318342] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951589, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.389317] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.425299] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951591, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.451347] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951593, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.818806] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951589, 'name': ReconfigVM_Task, 'duration_secs': 1.168821} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.819126] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1078.819857] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c484a67-e030-4e1f-aca4-f502af69ddbd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.842245] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 01708c16-43f8-408a-ae6b-82a1ad34eafa/01708c16-43f8-408a-ae6b-82a1ad34eafa.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.842496] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7be53b82-d3be-4aa0-8fbc-ace1522b6e6f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.859897] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1078.859897] env[61857]: value = "task-951594" [ 1078.859897] env[61857]: _type = "Task" [ 1078.859897] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.868322] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951594, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.887730] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.925239] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951591, 'name': PowerOffVM_Task, 'duration_secs': 1.163168} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.925527] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1078.925734] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance '04438fc0-b5a1-4e29-84d5-4d84d332c8e4' progress to 17 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1078.950535] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951593, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.369490] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951594, 'name': ReconfigVM_Task, 'duration_secs': 0.247568} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.369746] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 01708c16-43f8-408a-ae6b-82a1ad34eafa/01708c16-43f8-408a-ae6b-82a1ad34eafa.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1079.370019] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance '01708c16-43f8-408a-ae6b-82a1ad34eafa' progress to 50 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1079.389773] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.432633] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1079.432886] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1079.433063] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1079.433260] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1079.433412] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1079.433567] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1079.433780] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1079.433945] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1079.434141] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1079.434315] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1079.434497] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1079.439454] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8a2f401-48be-4450-8733-cb79216d5785 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.458816] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951593, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.460138] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1079.460138] env[61857]: value = "task-951595" [ 1079.460138] env[61857]: _type = "Task" [ 1079.460138] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.469601] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951595, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.877250] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d7c65c-d28d-4a23-ae8d-16e05c5c68a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.898130] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35caef6b-ac81-4fad-822d-5afef59aa7a6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.903248] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.917744] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance '01708c16-43f8-408a-ae6b-82a1ad34eafa' progress to 67 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1079.959400] env[61857]: DEBUG oslo_vmware.api [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951593, 'name': RemoveSnapshot_Task, 'duration_secs': 1.521544} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.959656] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1079.959885] env[61857]: INFO nova.compute.manager [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Took 14.74 seconds to snapshot the instance on the hypervisor. [ 1079.969957] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951595, 'name': ReconfigVM_Task, 'duration_secs': 0.186897} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.970245] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance '04438fc0-b5a1-4e29-84d5-4d84d332c8e4' progress to 33 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1080.391420] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.458615] env[61857]: DEBUG nova.network.neutron [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Port e1be624f-6a64-43aa-b429-788ac7363cc7 binding to destination host cpu-1 is already ACTIVE {{(pid=61857) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1080.476415] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.476661] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.476826] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.477163] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.477163] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.477323] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.477530] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.477690] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.477855] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.478032] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.478724] env[61857]: DEBUG nova.virt.hardware [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.483579] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1080.484177] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-271e9f46-1202-44e5-8f9b-92b485880778 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.504562] env[61857]: DEBUG nova.compute.manager [None req-5015b7a9-5990-4aaf-a754-d1f8e591b171 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Found 1 images (rotation: 2) {{(pid=61857) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1080.506840] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1080.506840] env[61857]: value = "task-951596" [ 1080.506840] env[61857]: _type = "Task" [ 1080.506840] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.515073] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951596, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.689576] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.689808] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.689958] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1080.891705] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.016509] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951596, 'name': ReconfigVM_Task, 'duration_secs': 0.155118} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.016761] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1081.017535] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2aab1f-138d-4dcb-9f58-3cac585540ed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.038899] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 04438fc0-b5a1-4e29-84d5-4d84d332c8e4/04438fc0-b5a1-4e29-84d5-4d84d332c8e4.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.039390] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76bbc01e-3195-4bfc-9765-a3c6ef74f3af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.056806] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1081.056806] env[61857]: value = "task-951597" [ 1081.056806] env[61857]: _type = "Task" [ 1081.056806] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.064636] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951597, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.161486] env[61857]: DEBUG nova.compute.manager [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1081.162498] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38186cbe-426c-43e1-a3a1-79c5efd6554c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.392432] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.476157] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "01708c16-43f8-408a-ae6b-82a1ad34eafa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.476412] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.476595] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.566344] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951597, 'name': ReconfigVM_Task, 'duration_secs': 0.23659} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.566627] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 04438fc0-b5a1-4e29-84d5-4d84d332c8e4/04438fc0-b5a1-4e29-84d5-4d84d332c8e4.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1081.566896] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance '04438fc0-b5a1-4e29-84d5-4d84d332c8e4' progress to 50 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1081.672955] env[61857]: INFO nova.compute.manager [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] instance snapshotting [ 1081.673615] env[61857]: DEBUG nova.objects.instance [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'flavor' on Instance uuid 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.893301] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.073186] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbdd5c5f-b3ec-4b1a-9554-ccb31c005d9f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.093534] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0975cc3-1ccc-4bd9-a749-ee5751eb5bc5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.112733] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance '04438fc0-b5a1-4e29-84d5-4d84d332c8e4' progress to 67 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1082.179424] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dfc1dbb-25e2-432e-b7bb-1b0ed18048cc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.198393] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b9654d-9d97-41cb-a998-a1d3a019419a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.201023] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Didn't find any instances for network info cache update. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1082.201225] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.201820] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.202054] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.202251] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.202407] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.202560] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.202690] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1082.202835] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.393710] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.514906] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.515118] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.515305] env[61857]: DEBUG nova.network.neutron [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1082.650981] env[61857]: DEBUG nova.network.neutron [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Port 83ff6e16-8b1b-4391-9ef1-57600502688c binding to destination host cpu-1 is already ACTIVE {{(pid=61857) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1082.709229] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.709458] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.709631] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.709787] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1082.710830] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1082.711696] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53046985-a7e8-4b0a-8481-db454e6842fe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.714707] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-869be817-c56e-4a95-9d4c-ea3d37740fd9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.721879] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd9b794-9955-4807-887f-014c6686ac29 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.726455] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1082.726455] env[61857]: value = "task-951598" [ 1082.726455] env[61857]: _type = "Task" [ 1082.726455] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.737918] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93f311f-a954-4fa3-bcb2-8e6488a8ca31 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.742730] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951598, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.746891] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f0ac99-1624-49fe-a7fd-bf2cf7eafe95 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.776517] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181133MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1082.776677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.776872] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.893995] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.218927] env[61857]: DEBUG nova.network.neutron [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance_info_cache with network_info: [{"id": "e1be624f-6a64-43aa-b429-788ac7363cc7", "address": "fa:16:3e:67:11:cc", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1be624f-6a", "ovs_interfaceid": "e1be624f-6a64-43aa-b429-788ac7363cc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.236879] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951598, 'name': CreateSnapshot_Task, 'duration_secs': 0.430613} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.237160] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1083.237881] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ff3bc8-b475-4a53-8f56-f54edc8a2202 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.394482] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.671667] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.671907] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.672105] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.721244] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.754319] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1083.754883] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0eb2931b-b9dc-4c9c-835e-e44feed9a06d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.763851] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1083.763851] env[61857]: value = "task-951599" [ 1083.763851] env[61857]: _type = "Task" [ 1083.763851] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.774125] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951599, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.784240] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Applying migration context for instance 01708c16-43f8-408a-ae6b-82a1ad34eafa as it has an incoming, in-progress migration b12e16cf-7bb8-4a79-bed3-ad040c5e8f5f. Migration status is post-migrating {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1083.784451] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Applying migration context for instance 04438fc0-b5a1-4e29-84d5-4d84d332c8e4 as it has an incoming, in-progress migration 75be059d-4418-4cfb-8c2d-feb19efd97a6. Migration status is post-migrating {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1083.785368] env[61857]: INFO nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating resource usage from migration b12e16cf-7bb8-4a79-bed3-ad040c5e8f5f [ 1083.785738] env[61857]: INFO nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating resource usage from migration 75be059d-4418-4cfb-8c2d-feb19efd97a6 [ 1083.804164] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 857dbf02-d829-4476-9094-13a30e14c799 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.804318] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 14844876-96ab-481d-980d-4133e89d7d4d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.804445] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance a42e8c77-3cc3-4323-b1ee-2a95192910d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.804570] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 14df609d-678c-4713-8c9c-f80fabb61c79 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.804714] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Migration b12e16cf-7bb8-4a79-bed3-ad040c5e8f5f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1083.804837] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 01708c16-43f8-408a-ae6b-82a1ad34eafa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.804954] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Migration 75be059d-4418-4cfb-8c2d-feb19efd97a6 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1083.805083] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 04438fc0-b5a1-4e29-84d5-4d84d332c8e4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.805262] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1083.805413] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2176MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1083.897539] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.917332] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3acbc99-9157-4c0a-9d8b-4ada2dcba0bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.925025] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58c8ce2-b1ee-4c0c-b61e-040f2b53b8af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.954394] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c5bcd8-4183-44d5-8b1e-ea71b1af4bf1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.961852] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c003355d-2af3-417c-9b56-f6948b65bce3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.977025] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.241362] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcca9dfa-ff27-4516-a885-52f90baa619b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.261514] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc366b8a-4f0b-4a81-ad44-ebcfe5083d99 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.269128] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance '01708c16-43f8-408a-ae6b-82a1ad34eafa' progress to 83 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1084.276961] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951599, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.395419] env[61857]: DEBUG oslo_vmware.api [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951590, 'name': ReconfigVM_Task, 'duration_secs': 6.78451} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.395637] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.395851] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Reconfigured VM to detach interface {{(pid=61857) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1084.480231] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.706151] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.706365] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.706548] env[61857]: DEBUG nova.network.neutron [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1084.777019] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1084.777019] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951599, 'name': CloneVM_Task, 'duration_secs': 0.914386} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.777019] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e96a4cc-1c5c-4374-ae9b-7ea89ecbbfa5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.778301] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Created linked-clone VM from snapshot [ 1084.779060] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8a3bc3-71c9-406b-a129-c88090590a1d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.787935] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Uploading image 61680048-5dc6-4f90-bce5-77295de497cf {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1084.790814] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1084.790814] env[61857]: value = "task-951600" [ 1084.790814] env[61857]: _type = "Task" [ 1084.790814] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.797921] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951600, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.808954] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1084.808954] env[61857]: value = "vm-214220" [ 1084.808954] env[61857]: _type = "VirtualMachine" [ 1084.808954] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1084.809209] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e36fead8-567c-4d67-90c3-e6d59203a9be {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.814492] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lease: (returnval){ [ 1084.814492] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52be982d-f599-90b4-2738-e82a13e69d8e" [ 1084.814492] env[61857]: _type = "HttpNfcLease" [ 1084.814492] env[61857]: } obtained for exporting VM: (result){ [ 1084.814492] env[61857]: value = "vm-214220" [ 1084.814492] env[61857]: _type = "VirtualMachine" [ 1084.814492] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1084.814787] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the lease: (returnval){ [ 1084.814787] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52be982d-f599-90b4-2738-e82a13e69d8e" [ 1084.814787] env[61857]: _type = "HttpNfcLease" [ 1084.814787] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1084.820509] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1084.820509] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52be982d-f599-90b4-2738-e82a13e69d8e" [ 1084.820509] env[61857]: _type = "HttpNfcLease" [ 1084.820509] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1084.986987] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1084.986987] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.209s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.986987] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1084.986987] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Cleaning up deleted instances with incomplete migration {{(pid=61857) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1085.304019] env[61857]: DEBUG oslo_vmware.api [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951600, 'name': PowerOnVM_Task, 'duration_secs': 0.398406} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.304019] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1085.304019] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-8657cce9-9184-4391-b594-3796ab2a65ec tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance '01708c16-43f8-408a-ae6b-82a1ad34eafa' progress to 100 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1085.322095] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1085.322095] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52be982d-f599-90b4-2738-e82a13e69d8e" [ 1085.322095] env[61857]: _type = "HttpNfcLease" [ 1085.322095] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1085.322748] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1085.322748] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52be982d-f599-90b4-2738-e82a13e69d8e" [ 1085.322748] env[61857]: _type = "HttpNfcLease" [ 1085.322748] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1085.325177] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59de08f8-b3cb-41d8-b5f2-61c617cbd2ad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.332740] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210d233-98e6-6aa5-a8c6-37732a1ddeaf/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1085.334718] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210d233-98e6-6aa5-a8c6-37732a1ddeaf/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1085.453840] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-28674ba6-930e-4495-a375-a7b0bb83bf60 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.488591] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1085.636037] env[61857]: DEBUG nova.network.neutron [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance_info_cache with network_info: [{"id": "83ff6e16-8b1b-4391-9ef1-57600502688c", "address": "fa:16:3e:46:c4:d7", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83ff6e16-8b", "ovs_interfaceid": "83ff6e16-8b1b-4391-9ef1-57600502688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.693891] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.694113] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.694305] env[61857]: DEBUG nova.network.neutron [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1085.887641] env[61857]: DEBUG nova.compute.manager [req-3e07c4e1-50dc-4ac4-a066-52fb699f6e32 req-84274c3e-f8f7-4142-bf43-69f674250d28 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received event network-changed-a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1085.887641] env[61857]: DEBUG nova.compute.manager [req-3e07c4e1-50dc-4ac4-a066-52fb699f6e32 req-84274c3e-f8f7-4142-bf43-69f674250d28 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing instance network info cache due to event network-changed-a5cec797-c084-4a3f-828a-2624cd347348. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1085.887641] env[61857]: DEBUG oslo_concurrency.lockutils [req-3e07c4e1-50dc-4ac4-a066-52fb699f6e32 req-84274c3e-f8f7-4142-bf43-69f674250d28 service nova] Acquiring lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.138738] env[61857]: DEBUG oslo_concurrency.lockutils [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.426670] env[61857]: INFO nova.network.neutron [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Port 5b14caa2-9c46-4c97-a303-7157978d0b97 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1086.427306] env[61857]: DEBUG nova.network.neutron [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updating instance_info_cache with network_info: [{"id": "a5cec797-c084-4a3f-828a-2624cd347348", "address": "fa:16:3e:e7:99:f4", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5cec797-c0", "ovs_interfaceid": "a5cec797-c084-4a3f-828a-2624cd347348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.555800] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "interface-14df609d-678c-4713-8c9c-f80fabb61c79-5b14caa2-9c46-4c97-a303-7157978d0b97" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.556210] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-14df609d-678c-4713-8c9c-f80fabb61c79-5b14caa2-9c46-4c97-a303-7157978d0b97" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.556667] env[61857]: DEBUG nova.objects.instance [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'flavor' on Instance uuid 14df609d-678c-4713-8c9c-f80fabb61c79 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.693810] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.730011] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa3746c-68b4-47ad-b7cc-f368712cab1f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.750611] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f30068d-de9d-4cd1-870c-7c0ac0d84f6d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.758766] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance '04438fc0-b5a1-4e29-84d5-4d84d332c8e4' progress to 83 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1086.931104] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.933481] env[61857]: DEBUG oslo_concurrency.lockutils [req-3e07c4e1-50dc-4ac4-a066-52fb699f6e32 req-84274c3e-f8f7-4142-bf43-69f674250d28 service nova] Acquired lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.933696] env[61857]: DEBUG nova.network.neutron [req-3e07c4e1-50dc-4ac4-a066-52fb699f6e32 req-84274c3e-f8f7-4142-bf43-69f674250d28 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Refreshing network info cache for port a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1087.162508] env[61857]: DEBUG nova.objects.instance [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'pci_requests' on Instance uuid 14df609d-678c-4713-8c9c-f80fabb61c79 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.199581] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1087.199581] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1087.199581] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Rebuilding the list of instances to heal {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1087.265429] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1087.265916] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23debadc-8ece-46a3-a33b-0a9c440f8ae7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.273933] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1087.273933] env[61857]: value = "task-951602" [ 1087.273933] env[61857]: _type = "Task" [ 1087.273933] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.282159] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951602, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.439420] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3b7a2d92-cc7e-4618-87d6-2f4187f20d04 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-14844876-96ab-481d-980d-4133e89d7d4d-5b14caa2-9c46-4c97-a303-7157978d0b97" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.641s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.651642] env[61857]: DEBUG nova.network.neutron [req-3e07c4e1-50dc-4ac4-a066-52fb699f6e32 req-84274c3e-f8f7-4142-bf43-69f674250d28 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updated VIF entry in instance network info cache for port a5cec797-c084-4a3f-828a-2624cd347348. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1087.652130] env[61857]: DEBUG nova.network.neutron [req-3e07c4e1-50dc-4ac4-a066-52fb699f6e32 req-84274c3e-f8f7-4142-bf43-69f674250d28 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updating instance_info_cache with network_info: [{"id": "a5cec797-c084-4a3f-828a-2624cd347348", "address": "fa:16:3e:e7:99:f4", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5cec797-c0", "ovs_interfaceid": "a5cec797-c084-4a3f-828a-2624cd347348", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.665600] env[61857]: DEBUG nova.objects.base [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Object Instance<14df609d-678c-4713-8c9c-f80fabb61c79> lazy-loaded attributes: flavor,pci_requests {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1087.665839] env[61857]: DEBUG nova.network.neutron [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1087.733608] env[61857]: DEBUG nova.policy [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9056d344ff614950a60983450a2a40ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5653ed6ab0a948ef89bb03cf7426dd10', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1087.737065] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.737237] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquired lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.737451] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Forcefully refreshing network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1087.737628] env[61857]: DEBUG nova.objects.instance [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lazy-loading 'info_cache' on Instance uuid a42e8c77-3cc3-4323-b1ee-2a95192910d5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.785148] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951602, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.028458] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "01708c16-43f8-408a-ae6b-82a1ad34eafa" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.028897] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.029296] env[61857]: DEBUG nova.compute.manager [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Going to confirm migration 4 {{(pid=61857) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1088.155708] env[61857]: DEBUG oslo_concurrency.lockutils [req-3e07c4e1-50dc-4ac4-a066-52fb699f6e32 req-84274c3e-f8f7-4142-bf43-69f674250d28 service nova] Releasing lock "refresh_cache-14844876-96ab-481d-980d-4133e89d7d4d" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.285520] env[61857]: DEBUG oslo_vmware.api [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951602, 'name': PowerOnVM_Task, 'duration_secs': 0.65869} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.285874] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1088.285987] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-2803892b-e19e-4b34-95b6-c6c877e33e07 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance '04438fc0-b5a1-4e29-84d5-4d84d332c8e4' progress to 100 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1088.587249] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.587473] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquired lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.587677] env[61857]: DEBUG nova.network.neutron [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1088.587890] env[61857]: DEBUG nova.objects.instance [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lazy-loading 'info_cache' on Instance uuid 01708c16-43f8-408a-ae6b-82a1ad34eafa {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.506830] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating instance_info_cache with network_info: [{"id": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "address": "fa:16:3e:52:2f:3f", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507983f7-c4", "ovs_interfaceid": "507983f7-c4c5-4605-887e-c3a1936e8f3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.718590] env[61857]: DEBUG nova.network.neutron [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Successfully updated port: 5b14caa2-9c46-4c97-a303-7157978d0b97 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1090.009772] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Releasing lock "refresh_cache-a42e8c77-3cc3-4323-b1ee-2a95192910d5" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.010030] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updated the network info_cache for instance {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1090.010384] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1090.010727] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1090.010727] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1090.010838] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1090.010953] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1090.016444] env[61857]: DEBUG nova.network.neutron [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance_info_cache with network_info: [{"id": "e1be624f-6a64-43aa-b429-788ac7363cc7", "address": "fa:16:3e:67:11:cc", "network": {"id": "915589f5-f890-4681-96e4-99ca7dde5954", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1180727509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62bd314daba846cd9ed3dc994dd1034b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f678cd81-6d15-43d5-aab7-d7eedc2ef2d5", "external-id": "nsx-vlan-transportzone-602", "segmentation_id": 602, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1be624f-6a", "ovs_interfaceid": "e1be624f-6a64-43aa-b429-788ac7363cc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.224741] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.224808] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.224976] env[61857]: DEBUG nova.network.neutron [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1090.514231] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.514587] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.514694] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.514803] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1090.515787] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dea8c6b-4777-4079-bd6c-0f6126a5f3ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.518976] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Releasing lock "refresh_cache-01708c16-43f8-408a-ae6b-82a1ad34eafa" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.519238] env[61857]: DEBUG nova.objects.instance [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lazy-loading 'migration_context' on Instance uuid 01708c16-43f8-408a-ae6b-82a1ad34eafa {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.525938] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1bc1a25-ce19-4153-b014-978fd6e25796 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.541241] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d646093-f395-4971-8f38-aa47377c6e36 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.548164] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9bb3c2-81a2-4680-af6c-d0cb4b25f02f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.579260] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181133MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1090.579442] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.579619] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.761633] env[61857]: WARNING nova.network.neutron [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] 30b7d27c-744c-46a6-8769-b1b1c95306e4 already exists in list: networks containing: ['30b7d27c-744c-46a6-8769-b1b1c95306e4']. ignoring it [ 1091.022141] env[61857]: DEBUG nova.objects.base [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Object Instance<01708c16-43f8-408a-ae6b-82a1ad34eafa> lazy-loaded attributes: info_cache,migration_context {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1091.023121] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c70481-f0a1-4a2f-905b-25773a37d9bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.043063] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1acb11fc-750a-47ba-ad5c-742ddf51bf16 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.049324] env[61857]: DEBUG oslo_vmware.api [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1091.049324] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]525c794f-27af-ae61-c2d0-73a5ab29ed39" [ 1091.049324] env[61857]: _type = "Task" [ 1091.049324] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.050534] env[61857]: DEBUG nova.network.neutron [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updating instance_info_cache with network_info: [{"id": "a8571bda-024e-4d91-9417-6b5f64325f87", "address": "fa:16:3e:e8:b5:c8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8571bda-02", "ovs_interfaceid": "a8571bda-024e-4d91-9417-6b5f64325f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5b14caa2-9c46-4c97-a303-7157978d0b97", "address": "fa:16:3e:17:98:ef", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b14caa2-9c", "ovs_interfaceid": "5b14caa2-9c46-4c97-a303-7157978d0b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.059915] env[61857]: DEBUG oslo_vmware.api [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]525c794f-27af-ae61-c2d0-73a5ab29ed39, 'name': SearchDatastore_Task, 'duration_secs': 0.008021} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.060220] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.556091] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.556830] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.557122] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.558415] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba97028-5951-44ea-826b-fd260518e427 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.576200] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.576472] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.576638] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.576836] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.576981] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.577237] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.577489] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.577663] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.577860] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.578050] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.578242] env[61857]: DEBUG nova.virt.hardware [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.584648] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Reconfiguring VM to attach interface {{(pid=61857) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1091.585047] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d49e19b-5466-4374-b6aa-26f4bca8837e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.597657] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Applying migration context for instance 01708c16-43f8-408a-ae6b-82a1ad34eafa as it has an incoming, in-progress migration b12e16cf-7bb8-4a79-bed3-ad040c5e8f5f. Migration status is finished {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1091.597895] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Applying migration context for instance 04438fc0-b5a1-4e29-84d5-4d84d332c8e4 as it has an incoming, in-progress migration 75be059d-4418-4cfb-8c2d-feb19efd97a6. Migration status is finished {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1091.598843] env[61857]: INFO nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating resource usage from migration b12e16cf-7bb8-4a79-bed3-ad040c5e8f5f [ 1091.599155] env[61857]: INFO nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating resource usage from migration 75be059d-4418-4cfb-8c2d-feb19efd97a6 [ 1091.607109] env[61857]: DEBUG oslo_vmware.api [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1091.607109] env[61857]: value = "task-951603" [ 1091.607109] env[61857]: _type = "Task" [ 1091.607109] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.616504] env[61857]: DEBUG oslo_vmware.api [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951603, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.620790] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 857dbf02-d829-4476-9094-13a30e14c799 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.620930] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 14844876-96ab-481d-980d-4133e89d7d4d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.621072] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance a42e8c77-3cc3-4323-b1ee-2a95192910d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.621194] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 14df609d-678c-4713-8c9c-f80fabb61c79 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.621339] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Migration b12e16cf-7bb8-4a79-bed3-ad040c5e8f5f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1091.621461] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 01708c16-43f8-408a-ae6b-82a1ad34eafa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.621577] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Migration 75be059d-4418-4cfb-8c2d-feb19efd97a6 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1091.621692] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 04438fc0-b5a1-4e29-84d5-4d84d332c8e4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.621882] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1091.622029] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2176MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1091.736382] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d71430d-6a12-4782-8350-51d07fa10ed2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.744444] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15dc7023-8aa4-4d93-a565-047dd7b0a107 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.776623] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa3f9c2-f4d0-4e6f-badd-837866f7abe2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.784360] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa7ec58-71a5-4dad-827c-af5111b2c8ce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.797701] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.117953] env[61857]: DEBUG oslo_vmware.api [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951603, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.300899] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.618878] env[61857]: DEBUG oslo_vmware.api [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951603, 'name': ReconfigVM_Task, 'duration_secs': 0.618228} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.619561] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.619870] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Reconfigured VM to attach interface {{(pid=61857) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1092.806140] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1092.806392] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.227s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.806669] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.746s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.807999] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.808167] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Cleaning up deleted instances {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1093.124624] env[61857]: DEBUG oslo_concurrency.lockutils [None req-45e9dad6-7240-4557-bccf-33e36a89f278 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-14df609d-678c-4713-8c9c-f80fabb61c79-5b14caa2-9c46-4c97-a303-7157978d0b97" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.568s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.330316] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] There are 57 instances to clean {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1093.330655] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 7ab9803b-8baa-4c49-b8df-61cd3b64b216] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1093.442758] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5fd0d83-99dc-46df-8286-babe349be4bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.450970] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bb45f4-36c7-40a8-9dff-1121d91aa465 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.482668] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b122a361-f1a2-4837-8e6e-d21b49cfba3f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.490394] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58811838-f51e-4302-8a03-31f35d1a9735 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.503730] env[61857]: DEBUG nova.compute.provider_tree [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.634852] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210d233-98e6-6aa5-a8c6-37732a1ddeaf/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1093.635782] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fed176-9e4a-493b-8040-5b68a7d9d83d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.642318] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210d233-98e6-6aa5-a8c6-37732a1ddeaf/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1093.642484] env[61857]: ERROR oslo_vmware.rw_handles [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210d233-98e6-6aa5-a8c6-37732a1ddeaf/disk-0.vmdk due to incomplete transfer. [ 1093.642725] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-66d2586a-fb01-4401-80ff-0dc04fc7d289 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.651074] env[61857]: DEBUG oslo_vmware.rw_handles [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210d233-98e6-6aa5-a8c6-37732a1ddeaf/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1093.651281] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Uploaded image 61680048-5dc6-4f90-bce5-77295de497cf to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1093.653838] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1093.654098] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bfc784e0-4338-42c9-a5d8-3041164284c2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.659924] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1093.659924] env[61857]: value = "task-951604" [ 1093.659924] env[61857]: _type = "Task" [ 1093.659924] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.667610] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951604, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.835143] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 5fff13ed-0914-4791-b518-ab0ab63fa52f] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1094.006532] env[61857]: DEBUG nova.scheduler.client.report [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1094.171621] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951604, 'name': Destroy_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.315187] env[61857]: DEBUG nova.compute.manager [req-6266e21a-46e4-48b9-9f91-e986e4dc453a req-81a89588-fb16-4620-bac3-7d261aca68e4 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Received event network-changed-a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1094.315450] env[61857]: DEBUG nova.compute.manager [req-6266e21a-46e4-48b9-9f91-e986e4dc453a req-81a89588-fb16-4620-bac3-7d261aca68e4 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Refreshing instance network info cache due to event network-changed-a8571bda-024e-4d91-9417-6b5f64325f87. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1094.315816] env[61857]: DEBUG oslo_concurrency.lockutils [req-6266e21a-46e4-48b9-9f91-e986e4dc453a req-81a89588-fb16-4620-bac3-7d261aca68e4 service nova] Acquiring lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.316071] env[61857]: DEBUG oslo_concurrency.lockutils [req-6266e21a-46e4-48b9-9f91-e986e4dc453a req-81a89588-fb16-4620-bac3-7d261aca68e4 service nova] Acquired lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.316344] env[61857]: DEBUG nova.network.neutron [req-6266e21a-46e4-48b9-9f91-e986e4dc453a req-81a89588-fb16-4620-bac3-7d261aca68e4 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Refreshing network info cache for port a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1094.337755] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 9bb676fa-a53a-419a-bc22-1ed618d7191f] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1094.672370] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951604, 'name': Destroy_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.840770] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 91bb8c20-b7b6-4994-8a0a-b91935397036] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.016406] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.210s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.032269] env[61857]: DEBUG nova.network.neutron [req-6266e21a-46e4-48b9-9f91-e986e4dc453a req-81a89588-fb16-4620-bac3-7d261aca68e4 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updated VIF entry in instance network info cache for port a8571bda-024e-4d91-9417-6b5f64325f87. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1095.032804] env[61857]: DEBUG nova.network.neutron [req-6266e21a-46e4-48b9-9f91-e986e4dc453a req-81a89588-fb16-4620-bac3-7d261aca68e4 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updating instance_info_cache with network_info: [{"id": "a8571bda-024e-4d91-9417-6b5f64325f87", "address": "fa:16:3e:e8:b5:c8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8571bda-02", "ovs_interfaceid": "a8571bda-024e-4d91-9417-6b5f64325f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5b14caa2-9c46-4c97-a303-7157978d0b97", "address": "fa:16:3e:17:98:ef", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b14caa2-9c", "ovs_interfaceid": "5b14caa2-9c46-4c97-a303-7157978d0b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.171050] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951604, 'name': Destroy_Task, 'duration_secs': 1.298279} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.171322] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Destroyed the VM [ 1095.171581] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1095.171836] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1aac0012-363f-49e8-adb5-ccd451500c86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.178706] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1095.178706] env[61857]: value = "task-951605" [ 1095.178706] env[61857]: _type = "Task" [ 1095.178706] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.187682] env[61857]: DEBUG oslo_concurrency.lockutils [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.187857] env[61857]: DEBUG oslo_concurrency.lockutils [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.192259] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951605, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.211891] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "interface-14df609d-678c-4713-8c9c-f80fabb61c79-5b14caa2-9c46-4c97-a303-7157978d0b97" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.212186] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-14df609d-678c-4713-8c9c-f80fabb61c79-5b14caa2-9c46-4c97-a303-7157978d0b97" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.223541] env[61857]: DEBUG nova.network.neutron [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Port 83ff6e16-8b1b-4391-9ef1-57600502688c binding to destination host cpu-1 is already ACTIVE {{(pid=61857) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1095.223811] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.223968] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.224172] env[61857]: DEBUG nova.network.neutron [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1095.344642] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 94720bbd-a5db-45ca-9cfc-02c0f127c8bc] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.536351] env[61857]: DEBUG oslo_concurrency.lockutils [req-6266e21a-46e4-48b9-9f91-e986e4dc453a req-81a89588-fb16-4620-bac3-7d261aca68e4 service nova] Releasing lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.574512] env[61857]: INFO nova.scheduler.client.report [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted allocation for migration b12e16cf-7bb8-4a79-bed3-ad040c5e8f5f [ 1095.692651] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951605, 'name': RemoveSnapshot_Task} progress is 16%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.693371] env[61857]: INFO nova.compute.manager [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Detaching volume f319926d-a413-4adf-b108-e133bfe9bd26 [ 1095.715256] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.715445] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.716294] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c66c29-02fd-4b5e-99b9-4c22ba227c59 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.736732] env[61857]: INFO nova.virt.block_device [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Attempting to driver detach volume f319926d-a413-4adf-b108-e133bfe9bd26 from mountpoint /dev/sdb [ 1095.736961] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Volume detach. Driver type: vmdk {{(pid=61857) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1095.737167] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214206', 'volume_id': 'f319926d-a413-4adf-b108-e133bfe9bd26', 'name': 'volume-f319926d-a413-4adf-b108-e133bfe9bd26', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'a42e8c77-3cc3-4323-b1ee-2a95192910d5', 'attached_at': '', 'detached_at': '', 'volume_id': 'f319926d-a413-4adf-b108-e133bfe9bd26', 'serial': 'f319926d-a413-4adf-b108-e133bfe9bd26'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1095.737935] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63ffa2a-182b-458b-b366-4f46c7471445 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.740788] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9cf1dc-0732-41a0-a9ff-977ba90b12ab {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.776137] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b70403-568b-406d-9960-1ef779717188 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.785591] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Reconfiguring VM to detach interface {{(pid=61857) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1095.787793] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9582dbb1-ad2a-4545-93fa-3a49766eebdc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.803816] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5250349-b4b8-4555-80fd-7923854d63f8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.806865] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1095.806865] env[61857]: value = "task-951606" [ 1095.806865] env[61857]: _type = "Task" [ 1095.806865] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.825664] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d958b5f8-0335-4327-b998-07fd3f472cf6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.831311] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.843618] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] The volume has not been displaced from its original location: [datastore1] volume-f319926d-a413-4adf-b108-e133bfe9bd26/volume-f319926d-a413-4adf-b108-e133bfe9bd26.vmdk. No consolidation needed. {{(pid=61857) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1095.848821] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Reconfiguring VM instance instance-00000059 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1095.851472] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e2c3acbd-e7fd-45c7-8e99-ef9b87fb106c] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.853157] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37fcbecb-36de-4f5d-ade3-774a3d590b03 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.871455] env[61857]: DEBUG oslo_vmware.api [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1095.871455] env[61857]: value = "task-951607" [ 1095.871455] env[61857]: _type = "Task" [ 1095.871455] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.879473] env[61857]: DEBUG oslo_vmware.api [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951607, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.015417] env[61857]: DEBUG nova.network.neutron [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance_info_cache with network_info: [{"id": "83ff6e16-8b1b-4391-9ef1-57600502688c", "address": "fa:16:3e:46:c4:d7", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83ff6e16-8b", "ovs_interfaceid": "83ff6e16-8b1b-4391-9ef1-57600502688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.080147] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.051s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.190532] env[61857]: DEBUG oslo_vmware.api [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951605, 'name': RemoveSnapshot_Task, 'duration_secs': 0.535284} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.190869] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1096.191156] env[61857]: INFO nova.compute.manager [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Took 14.01 seconds to snapshot the instance on the hypervisor. [ 1096.317283] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.340993] env[61857]: DEBUG nova.compute.manager [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Received event network-vif-plugged-5b14caa2-9c46-4c97-a303-7157978d0b97 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1096.341232] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] Acquiring lock "14df609d-678c-4713-8c9c-f80fabb61c79-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.341456] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] Lock "14df609d-678c-4713-8c9c-f80fabb61c79-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.341622] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] Lock "14df609d-678c-4713-8c9c-f80fabb61c79-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.341794] env[61857]: DEBUG nova.compute.manager [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] No waiting events found dispatching network-vif-plugged-5b14caa2-9c46-4c97-a303-7157978d0b97 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1096.341965] env[61857]: WARNING nova.compute.manager [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Received unexpected event network-vif-plugged-5b14caa2-9c46-4c97-a303-7157978d0b97 for instance with vm_state active and task_state None. [ 1096.342143] env[61857]: DEBUG nova.compute.manager [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Received event network-changed-5b14caa2-9c46-4c97-a303-7157978d0b97 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1096.342305] env[61857]: DEBUG nova.compute.manager [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Refreshing instance network info cache due to event network-changed-5b14caa2-9c46-4c97-a303-7157978d0b97. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1096.342489] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] Acquiring lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.342647] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] Acquired lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.342829] env[61857]: DEBUG nova.network.neutron [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Refreshing network info cache for port 5b14caa2-9c46-4c97-a303-7157978d0b97 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1096.366666] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: fb993549-1638-4aa5-bee3-9f303a7e5cfd] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1096.382862] env[61857]: DEBUG oslo_vmware.api [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951607, 'name': ReconfigVM_Task, 'duration_secs': 0.22963} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.383139] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Reconfigured VM instance instance-00000059 to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1096.387879] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68a58e06-4f0f-48f6-9435-c9f9f170ccf4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.402668] env[61857]: DEBUG oslo_vmware.api [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1096.402668] env[61857]: value = "task-951608" [ 1096.402668] env[61857]: _type = "Task" [ 1096.402668] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.410669] env[61857]: DEBUG oslo_vmware.api [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951608, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.518438] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.748491] env[61857]: DEBUG nova.compute.manager [None req-c6a27a79-88f7-4bc0-9d3c-31e21d99ed28 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Found 2 images (rotation: 2) {{(pid=61857) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1096.817330] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.869180] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 2cd4f38e-4111-45b2-a8bd-dc0d38ada293] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1096.911975] env[61857]: DEBUG oslo_vmware.api [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951608, 'name': ReconfigVM_Task, 'duration_secs': 0.143565} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.914319] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214206', 'volume_id': 'f319926d-a413-4adf-b108-e133bfe9bd26', 'name': 'volume-f319926d-a413-4adf-b108-e133bfe9bd26', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'a42e8c77-3cc3-4323-b1ee-2a95192910d5', 'attached_at': '', 'detached_at': '', 'volume_id': 'f319926d-a413-4adf-b108-e133bfe9bd26', 'serial': 'f319926d-a413-4adf-b108-e133bfe9bd26'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1097.023476] env[61857]: DEBUG nova.compute.manager [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61857) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1097.023476] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.023476] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.052611] env[61857]: DEBUG nova.network.neutron [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updated VIF entry in instance network info cache for port 5b14caa2-9c46-4c97-a303-7157978d0b97. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1097.052611] env[61857]: DEBUG nova.network.neutron [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updating instance_info_cache with network_info: [{"id": "a8571bda-024e-4d91-9417-6b5f64325f87", "address": "fa:16:3e:e8:b5:c8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8571bda-02", "ovs_interfaceid": "a8571bda-024e-4d91-9417-6b5f64325f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5b14caa2-9c46-4c97-a303-7157978d0b97", "address": "fa:16:3e:17:98:ef", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b14caa2-9c", "ovs_interfaceid": "5b14caa2-9c46-4c97-a303-7157978d0b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.318703] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.372858] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: ce67a15a-8604-4523-a8fa-a34ccf9914da] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1097.459337] env[61857]: DEBUG nova.objects.instance [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lazy-loading 'flavor' on Instance uuid a42e8c77-3cc3-4323-b1ee-2a95192910d5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.524971] env[61857]: DEBUG nova.objects.instance [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'migration_context' on Instance uuid 04438fc0-b5a1-4e29-84d5-4d84d332c8e4 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.543362] env[61857]: DEBUG nova.compute.manager [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1097.544337] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e265e7-4677-4ef5-a180-6c2421d1a53d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.551204] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "01708c16-43f8-408a-ae6b-82a1ad34eafa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.551548] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.551890] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "01708c16-43f8-408a-ae6b-82a1ad34eafa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.552162] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.552404] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.559022] env[61857]: INFO nova.compute.manager [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Terminating instance [ 1097.559022] env[61857]: DEBUG oslo_concurrency.lockutils [req-bf3c178f-a4ef-4878-b691-7c41cd5b7a80 req-5d786de6-467c-427c-bf68-7c416734395c service nova] Releasing lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.559897] env[61857]: DEBUG nova.compute.manager [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1097.560105] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1097.560866] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a18e460-bd9f-45e0-9d34-227c6796f887 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.570616] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1097.570845] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d475947-4cdc-4537-accb-0903af8b7365 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.576425] env[61857]: DEBUG oslo_vmware.api [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1097.576425] env[61857]: value = "task-951609" [ 1097.576425] env[61857]: _type = "Task" [ 1097.576425] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.585461] env[61857]: DEBUG oslo_vmware.api [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951609, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.819253] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.876321] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: fd43f45b-0f54-46c6-94f9-ce28d3b15bb1] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.060287] env[61857]: INFO nova.compute.manager [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] instance snapshotting [ 1098.060878] env[61857]: DEBUG nova.objects.instance [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'flavor' on Instance uuid 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.088169] env[61857]: DEBUG oslo_vmware.api [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951609, 'name': PowerOffVM_Task, 'duration_secs': 0.147684} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.088441] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1098.088612] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1098.088864] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e267c260-7e17-4f18-a534-b1180401bbf0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.122891] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906ea65c-adfb-4847-a66d-e75ed939bfe5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.129793] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71dbd41-502d-4690-ad24-a2c716dac606 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.160758] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913d1fed-e1ad-406f-ad14-6e518b5374a2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.163321] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1098.163522] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1098.163702] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleting the datastore file [datastore1] 01708c16-43f8-408a-ae6b-82a1ad34eafa {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.163937] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0be92e08-73c8-4cfd-8a47-6ca4990c1ad4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.170504] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5d983c-3b4c-4ff7-b80e-426be1b2b17f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.175677] env[61857]: DEBUG oslo_vmware.api [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for the task: (returnval){ [ 1098.175677] env[61857]: value = "task-951611" [ 1098.175677] env[61857]: _type = "Task" [ 1098.175677] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.189854] env[61857]: DEBUG nova.compute.provider_tree [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.195841] env[61857]: DEBUG oslo_vmware.api [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951611, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.320313] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.380025] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c821ec43-9462-4276-bdc9-d54e64c0a3cf] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.467184] env[61857]: DEBUG oslo_concurrency.lockutils [None req-20131b51-b215-465c-9cbf-3f4f59065d1b tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.279s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.565703] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6bc03e-8754-40ca-8ba2-69a29dfadb28 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.585016] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e1fdcc-1806-484c-b382-931ac43667d4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.685422] env[61857]: DEBUG oslo_vmware.api [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Task: {'id': task-951611, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141267} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.685967] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.686190] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1098.686382] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1098.686564] env[61857]: INFO nova.compute.manager [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1098.686808] env[61857]: DEBUG oslo.service.loopingcall [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.687016] env[61857]: DEBUG nova.compute.manager [-] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1098.687117] env[61857]: DEBUG nova.network.neutron [-] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1098.692718] env[61857]: DEBUG nova.scheduler.client.report [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1098.711785] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.712167] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.712495] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.712740] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.712916] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.714778] env[61857]: INFO nova.compute.manager [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Terminating instance [ 1098.718097] env[61857]: DEBUG nova.compute.manager [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1098.718317] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1098.719573] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de66bd4-9be8-4bac-9cac-00a352d15e97 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.728058] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1098.728302] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4124b89-2c2d-4324-842e-d309f85b2489 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.734013] env[61857]: DEBUG oslo_vmware.api [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1098.734013] env[61857]: value = "task-951612" [ 1098.734013] env[61857]: _type = "Task" [ 1098.734013] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.742046] env[61857]: DEBUG oslo_vmware.api [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951612, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.822787] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.883973] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: ce5f713d-8b63-4b95-a5ad-feddb95f921d] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.932972] env[61857]: DEBUG nova.compute.manager [req-a80c680d-81cf-432d-9123-d5a907913210 req-b73823b4-0a05-4556-afe2-5a04636fb9fd service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Received event network-vif-deleted-e1be624f-6a64-43aa-b429-788ac7363cc7 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1098.933211] env[61857]: INFO nova.compute.manager [req-a80c680d-81cf-432d-9123-d5a907913210 req-b73823b4-0a05-4556-afe2-5a04636fb9fd service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Neutron deleted interface e1be624f-6a64-43aa-b429-788ac7363cc7; detaching it from the instance and deleting it from the info cache [ 1098.933393] env[61857]: DEBUG nova.network.neutron [req-a80c680d-81cf-432d-9123-d5a907913210 req-b73823b4-0a05-4556-afe2-5a04636fb9fd service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.094854] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1099.095162] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b8a99657-6bee-4196-bf2b-2b2e08fd835b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.102274] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1099.102274] env[61857]: value = "task-951613" [ 1099.102274] env[61857]: _type = "Task" [ 1099.102274] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.109854] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951613, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.243173] env[61857]: DEBUG oslo_vmware.api [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951612, 'name': PowerOffVM_Task, 'duration_secs': 0.190165} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.243451] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1099.243624] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1099.243879] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9810d3b6-5563-4453-9107-f93fba712dfb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.306095] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1099.306354] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1099.306545] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleting the datastore file [datastore2] a42e8c77-3cc3-4323-b1ee-2a95192910d5 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1099.306823] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7137ad4d-e16f-4ec3-a321-e652aa1bc6ad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.313298] env[61857]: DEBUG oslo_vmware.api [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1099.313298] env[61857]: value = "task-951615" [ 1099.313298] env[61857]: _type = "Task" [ 1099.313298] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.323871] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.326658] env[61857]: DEBUG oslo_vmware.api [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951615, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.387024] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 825f5d87-76b1-4e4e-9b3d-4be419c7d323] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1099.409506] env[61857]: DEBUG nova.network.neutron [-] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.436096] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e450c761-68fc-4a44-83bd-ac69fc032794 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.445161] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2822a21-41a7-436b-a2cd-8d9dbf15e847 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.473300] env[61857]: DEBUG nova.compute.manager [req-a80c680d-81cf-432d-9123-d5a907913210 req-b73823b4-0a05-4556-afe2-5a04636fb9fd service nova] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Detach interface failed, port_id=e1be624f-6a64-43aa-b429-788ac7363cc7, reason: Instance 01708c16-43f8-408a-ae6b-82a1ad34eafa could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1099.613786] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951613, 'name': CreateSnapshot_Task, 'duration_secs': 0.415654} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.614038] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1099.614769] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5d4fec-d9e5-4cb3-a9c7-d25bb5a43ce6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.703165] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.680s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.825939] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.828844] env[61857]: DEBUG oslo_vmware.api [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951615, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125346} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.829101] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1099.829304] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1099.829518] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1099.829705] env[61857]: INFO nova.compute.manager [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1099.829948] env[61857]: DEBUG oslo.service.loopingcall [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1099.830158] env[61857]: DEBUG nova.compute.manager [-] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1099.830250] env[61857]: DEBUG nova.network.neutron [-] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1099.890454] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: d85f8a2d-6053-45ec-81bb-de9fe02af3fe] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1099.912580] env[61857]: INFO nova.compute.manager [-] [instance: 01708c16-43f8-408a-ae6b-82a1ad34eafa] Took 1.23 seconds to deallocate network for instance. [ 1100.131495] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1100.131826] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-737cdfe9-b5b3-4fd3-b208-4a65502fc518 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.141032] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1100.141032] env[61857]: value = "task-951616" [ 1100.141032] env[61857]: _type = "Task" [ 1100.141032] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.148443] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951616, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.324076] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.393449] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: f982f0c5-c701-4e5b-b96d-c21a3c0da24f] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1100.420386] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.420791] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.421017] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.446900] env[61857]: INFO nova.scheduler.client.report [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Deleted allocations for instance 01708c16-43f8-408a-ae6b-82a1ad34eafa [ 1100.652364] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951616, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.733984] env[61857]: DEBUG nova.network.neutron [-] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.824618] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.897012] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 992bcab3-7396-495f-96a1-e99ecd13c961] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1100.959732] env[61857]: DEBUG nova.compute.manager [req-47331e2d-04c7-47f1-a30d-f1d252669222 req-779e081a-92d3-4a2b-9314-e8acc6f271c4 service nova] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Received event network-vif-deleted-507983f7-c4c5-4605-887e-c3a1936e8f3a {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1100.962608] env[61857]: DEBUG oslo_concurrency.lockutils [None req-e4af91e9-26cb-4472-babf-bd9556250c16 tempest-DeleteServersTestJSON-1548661960 tempest-DeleteServersTestJSON-1548661960-project-member] Lock "01708c16-43f8-408a-ae6b-82a1ad34eafa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.411s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.151206] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951616, 'name': CloneVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.235683] env[61857]: INFO nova.compute.manager [-] [instance: a42e8c77-3cc3-4323-b1ee-2a95192910d5] Took 1.41 seconds to deallocate network for instance. [ 1101.242412] env[61857]: INFO nova.compute.manager [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Swapping old allocation on dict_keys(['2d5860d0-8574-4e55-9ced-20e66f0314c2']) held by migration 75be059d-4418-4cfb-8c2d-feb19efd97a6 for instance [ 1101.263633] env[61857]: DEBUG nova.scheduler.client.report [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Overwriting current allocation {'allocations': {'2d5860d0-8574-4e55-9ced-20e66f0314c2': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 137}}, 'project_id': '5ef111c8b3ff4f2383b4e81db026792b', 'user_id': 'f719cbf1755d4c34bef92dab6679cb56', 'consumer_generation': 1} on consumer 04438fc0-b5a1-4e29-84d5-4d84d332c8e4 {{(pid=61857) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1101.325110] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.340694] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.340909] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.341107] env[61857]: DEBUG nova.network.neutron [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1101.400601] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 18a632be-322f-48a9-becf-51ff3b735ad4] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1101.651495] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951616, 'name': CloneVM_Task, 'duration_secs': 1.082091} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.651904] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Created linked-clone VM from snapshot [ 1101.653029] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bbcda6-f604-4896-b819-b10f0f3cf670 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.663176] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Uploading image 788e8d57-21f0-4405-a84a-f5f5a69c06d1 {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1101.689707] env[61857]: DEBUG oslo_vmware.rw_handles [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1101.689707] env[61857]: value = "vm-214222" [ 1101.689707] env[61857]: _type = "VirtualMachine" [ 1101.689707] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1101.690087] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-07f2af34-394b-4f8c-89b5-797883cda9d1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.697732] env[61857]: DEBUG oslo_vmware.rw_handles [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lease: (returnval){ [ 1101.697732] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52549a9c-ce79-f01d-e1c0-21112731f00a" [ 1101.697732] env[61857]: _type = "HttpNfcLease" [ 1101.697732] env[61857]: } obtained for exporting VM: (result){ [ 1101.697732] env[61857]: value = "vm-214222" [ 1101.697732] env[61857]: _type = "VirtualMachine" [ 1101.697732] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1101.698029] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the lease: (returnval){ [ 1101.698029] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52549a9c-ce79-f01d-e1c0-21112731f00a" [ 1101.698029] env[61857]: _type = "HttpNfcLease" [ 1101.698029] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1101.704123] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1101.704123] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52549a9c-ce79-f01d-e1c0-21112731f00a" [ 1101.704123] env[61857]: _type = "HttpNfcLease" [ 1101.704123] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1101.742328] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.742623] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.742870] env[61857]: DEBUG nova.objects.instance [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lazy-loading 'resources' on Instance uuid a42e8c77-3cc3-4323-b1ee-2a95192910d5 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.824920] env[61857]: DEBUG oslo_vmware.api [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951606, 'name': ReconfigVM_Task, 'duration_secs': 5.73968} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.825706] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.825706] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Reconfigured VM to detach interface {{(pid=61857) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1101.903774] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 0de8e67e-24b9-48d0-ac90-b5ec5b93e243] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1102.070444] env[61857]: DEBUG nova.network.neutron [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance_info_cache with network_info: [{"id": "83ff6e16-8b1b-4391-9ef1-57600502688c", "address": "fa:16:3e:46:c4:d7", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83ff6e16-8b", "ovs_interfaceid": "83ff6e16-8b1b-4391-9ef1-57600502688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.206039] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1102.206039] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52549a9c-ce79-f01d-e1c0-21112731f00a" [ 1102.206039] env[61857]: _type = "HttpNfcLease" [ 1102.206039] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1102.206350] env[61857]: DEBUG oslo_vmware.rw_handles [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1102.206350] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52549a9c-ce79-f01d-e1c0-21112731f00a" [ 1102.206350] env[61857]: _type = "HttpNfcLease" [ 1102.206350] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1102.207059] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720a0922-9f31-44e9-a311-bdd38605d36a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.214020] env[61857]: DEBUG oslo_vmware.rw_handles [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f77343-a6bd-6c0d-f45d-c0eef81c786c/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1102.214209] env[61857]: DEBUG oslo_vmware.rw_handles [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f77343-a6bd-6c0d-f45d-c0eef81c786c/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1102.303456] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-528e215c-e216-445b-bfce-c33b4afe332d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.345848] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b136fca3-4937-4911-80f0-f64324488dd3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.352867] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1376b892-b7cf-4c5a-9d48-3b8089a282c3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.021124] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: d979a8a1-a5b5-428e-9b60-7d827337218f] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.022996] env[61857]: DEBUG oslo_concurrency.lockutils [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-04438fc0-b5a1-4e29-84d5-4d84d332c8e4" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.023410] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1103.025200] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9487c2f7-5c32-4751-9afd-e8d07e0e9f78 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.058685] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35406ef2-8f33-4fa7-83f1-ee28e7dcc3fe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.061459] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1103.061459] env[61857]: value = "task-951619" [ 1103.061459] env[61857]: _type = "Task" [ 1103.061459] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.069317] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec42ecb-dc22-414e-9781-67d08d9bf5a2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.076450] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.087013] env[61857]: DEBUG nova.compute.provider_tree [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.525580] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 33cb5bbc-926d-42ee-b483-8d1e24707e40] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.527931] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.528203] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquired lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.528757] env[61857]: DEBUG nova.network.neutron [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1103.574327] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951619, 'name': PowerOffVM_Task, 'duration_secs': 0.196691} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.574672] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1103.575345] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1103.575707] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1103.576009] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.576270] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1103.576568] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.576793] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1103.577070] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1103.577290] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1103.577503] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1103.577767] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1103.577976] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.583067] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd1cd51f-37ef-45a0-a50e-2f29624d0979 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.594371] env[61857]: DEBUG nova.scheduler.client.report [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1103.602446] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1103.602446] env[61857]: value = "task-951620" [ 1103.602446] env[61857]: _type = "Task" [ 1103.602446] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.611050] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951620, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.634965] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "14df609d-678c-4713-8c9c-f80fabb61c79" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.635317] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14df609d-678c-4713-8c9c-f80fabb61c79" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.635568] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "14df609d-678c-4713-8c9c-f80fabb61c79-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.635819] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14df609d-678c-4713-8c9c-f80fabb61c79-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.636064] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14df609d-678c-4713-8c9c-f80fabb61c79-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.638356] env[61857]: INFO nova.compute.manager [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Terminating instance [ 1103.640492] env[61857]: DEBUG nova.compute.manager [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1103.640841] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1103.641781] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4a47b8-d5dc-4d5e-abab-0a1d0250d52b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.649951] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1103.651619] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3284688-3e9a-48b0-92d5-2c84411797c2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.656387] env[61857]: DEBUG oslo_vmware.api [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1103.656387] env[61857]: value = "task-951621" [ 1103.656387] env[61857]: _type = "Task" [ 1103.656387] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.665048] env[61857]: DEBUG oslo_vmware.api [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951621, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.031377] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 0ea7c119-f06d-4cbb-9b0d-77f3d11cb9cc] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1104.099794] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.357s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.113458] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951620, 'name': ReconfigVM_Task, 'duration_secs': 0.131432} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.114566] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69340ad4-4aaf-44ed-90c7-7866bcbce53a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.133905] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1104.134306] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1104.134494] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1104.134737] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1104.134943] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1104.135142] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1104.135461] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1104.135794] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1104.136068] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1104.136366] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1104.136591] env[61857]: DEBUG nova.virt.hardware [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1104.138211] env[61857]: INFO nova.scheduler.client.report [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleted allocations for instance a42e8c77-3cc3-4323-b1ee-2a95192910d5 [ 1104.139579] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2751bbf-023f-46b2-8f9b-58fe5a774e61 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.150370] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1104.150370] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]520489f2-7cea-03ae-2710-1d8c63d4a49b" [ 1104.150370] env[61857]: _type = "Task" [ 1104.150370] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.159256] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]520489f2-7cea-03ae-2710-1d8c63d4a49b, 'name': SearchDatastore_Task, 'duration_secs': 0.006885} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.167567] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1104.168315] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-529ef2b9-1a73-46b2-9c84-8f405a1bbdcb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.186914] env[61857]: DEBUG oslo_vmware.api [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951621, 'name': PowerOffVM_Task, 'duration_secs': 0.261133} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.189470] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1104.189470] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1104.189470] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1104.189470] env[61857]: value = "task-951622" [ 1104.189470] env[61857]: _type = "Task" [ 1104.189470] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.189470] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e05fa92d-0627-4bd9-8df6-3ab4538aa852 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.201345] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951622, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.261837] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1104.262177] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1104.262385] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleting the datastore file [datastore2] 14df609d-678c-4713-8c9c-f80fabb61c79 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1104.263126] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da31ae82-7a05-40fd-a1c7-87e234881621 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.270054] env[61857]: DEBUG oslo_vmware.api [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1104.270054] env[61857]: value = "task-951624" [ 1104.270054] env[61857]: _type = "Task" [ 1104.270054] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.279738] env[61857]: DEBUG oslo_vmware.api [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951624, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.465576] env[61857]: INFO nova.network.neutron [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Port 5b14caa2-9c46-4c97-a303-7157978d0b97 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1104.465947] env[61857]: DEBUG nova.network.neutron [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updating instance_info_cache with network_info: [{"id": "a8571bda-024e-4d91-9417-6b5f64325f87", "address": "fa:16:3e:e8:b5:c8", "network": {"id": "30b7d27c-744c-46a6-8769-b1b1c95306e4", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1307102604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5653ed6ab0a948ef89bb03cf7426dd10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8571bda-02", "ovs_interfaceid": "a8571bda-024e-4d91-9417-6b5f64325f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.538277] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: d9d05629-79ef-45c6-ac54-b3bc18d306f4] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1104.650259] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9473f9b7-0cc3-4d2b-b28d-34b4b1c567c9 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "a42e8c77-3cc3-4323-b1ee-2a95192910d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.938s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.705639] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951622, 'name': ReconfigVM_Task, 'duration_secs': 0.226762} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.705639] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1104.706615] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-094ae955-3c96-4054-a045-ca0da2502dc2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.743985] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 04438fc0-b5a1-4e29-84d5-4d84d332c8e4/04438fc0-b5a1-4e29-84d5-4d84d332c8e4.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.745198] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4c17e69-5ddd-4513-af28-8e9ae2350cf4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.778738] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1104.778738] env[61857]: value = "task-951625" [ 1104.778738] env[61857]: _type = "Task" [ 1104.778738] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.789088] env[61857]: DEBUG oslo_vmware.api [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951624, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189694} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.790133] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1104.790425] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1104.790720] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1104.790963] env[61857]: INFO nova.compute.manager [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1104.791350] env[61857]: DEBUG oslo.service.loopingcall [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1104.796693] env[61857]: DEBUG nova.compute.manager [-] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1104.796917] env[61857]: DEBUG nova.network.neutron [-] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1104.799677] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951625, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.972021] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Releasing lock "refresh_cache-14df609d-678c-4713-8c9c-f80fabb61c79" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.049418] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 6957dc98-8c60-4fdd-83dd-be2f13825c6d] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1105.294111] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951625, 'name': ReconfigVM_Task, 'duration_secs': 0.423128} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.294111] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 04438fc0-b5a1-4e29-84d5-4d84d332c8e4/04438fc0-b5a1-4e29-84d5-4d84d332c8e4.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.296034] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae625da0-c2d8-4718-b0b5-528bf889b006 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.331331] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2628e8-e336-44be-9e03-7525a73467d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.361912] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086104a3-7472-4c0d-a54e-79296bcc48d3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.382906] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f01fa8c-43fa-4b58-888d-d93ad97d6f59 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.390413] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1105.390641] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b0adf2b-b0e0-4442-b662-a7416ed01799 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.397046] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1105.397046] env[61857]: value = "task-951626" [ 1105.397046] env[61857]: _type = "Task" [ 1105.397046] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.406144] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951626, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.476646] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5a963ce9-b2d2-4f92-8054-2fb6148e6553 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "interface-14df609d-678c-4713-8c9c-f80fabb61c79-5b14caa2-9c46-4c97-a303-7157978d0b97" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.264s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.551987] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: ffa63de7-dd32-4908-92c5-755b3c0799ef] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1105.907704] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951626, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.056105] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: ddeed65b-9003-443f-8b2b-0756fbe7d234] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1106.413600] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951626, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.426468] env[61857]: DEBUG nova.compute.manager [req-a879aed6-2dff-4a9a-9ea7-f7032964a971 req-6ce3a029-d924-41a8-83f9-4d0488f8bb73 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Received event network-vif-deleted-a8571bda-024e-4d91-9417-6b5f64325f87 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1106.426468] env[61857]: INFO nova.compute.manager [req-a879aed6-2dff-4a9a-9ea7-f7032964a971 req-6ce3a029-d924-41a8-83f9-4d0488f8bb73 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Neutron deleted interface a8571bda-024e-4d91-9417-6b5f64325f87; detaching it from the instance and deleting it from the info cache [ 1106.426468] env[61857]: DEBUG nova.network.neutron [req-a879aed6-2dff-4a9a-9ea7-f7032964a971 req-6ce3a029-d924-41a8-83f9-4d0488f8bb73 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.562965] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 3fb5d24b-1767-43d9-bed5-833121962adb] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1106.895177] env[61857]: DEBUG nova.network.neutron [-] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.908922] env[61857]: DEBUG oslo_vmware.api [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951626, 'name': PowerOnVM_Task, 'duration_secs': 1.131292} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.909846] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1106.930813] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c7496e1-c00d-4453-8118-2786d151397a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.939022] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec29d964-1c5e-4297-8037-8699fd2d1693 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.966016] env[61857]: DEBUG nova.compute.manager [req-a879aed6-2dff-4a9a-9ea7-f7032964a971 req-6ce3a029-d924-41a8-83f9-4d0488f8bb73 service nova] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Detach interface failed, port_id=a8571bda-024e-4d91-9417-6b5f64325f87, reason: Instance 14df609d-678c-4713-8c9c-f80fabb61c79 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1107.065960] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c27f6f5b-6423-4eee-be53-dc48c93df893] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1107.404069] env[61857]: INFO nova.compute.manager [-] [instance: 14df609d-678c-4713-8c9c-f80fabb61c79] Took 2.61 seconds to deallocate network for instance. [ 1107.569340] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 7a9252da-2584-40ed-9d28-ca7341ed5165] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1107.575368] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "c6d20166-7251-406f-9003-b0599c2f5987" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.576280] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "c6d20166-7251-406f-9003-b0599c2f5987" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.912174] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.913054] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.913054] env[61857]: DEBUG nova.objects.instance [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'resources' on Instance uuid 14df609d-678c-4713-8c9c-f80fabb61c79 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.919407] env[61857]: INFO nova.compute.manager [None req-8315d6d4-8055-47a7-a2d5-7fb36b9fd9f3 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance to original state: 'active' [ 1108.076062] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: ddd4c436-405f-49f7-8c9b-de3b71725f63] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1108.079758] env[61857]: DEBUG nova.compute.manager [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1108.503017] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4441cd0d-79be-4512-af96-d7de52f86c51 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.510690] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1155dd5e-310d-436c-8fa3-e53c6b9add96 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.545668] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c945f8d-6589-462c-8377-4d89c31d6988 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.556021] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534f4d52-a3cc-4b5e-bf78-9216d6fb1233 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.572820] env[61857]: DEBUG nova.compute.provider_tree [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.583643] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: d29f6e34-861d-4cbf-8cbd-ab193a55220b] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1108.607640] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.077547] env[61857]: DEBUG nova.scheduler.client.report [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1109.090740] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: ad89cb5f-44d0-45c9-92a5-7f7ff3138e21] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1109.582779] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.670s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.585252] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.978s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.586828] env[61857]: INFO nova.compute.claims [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1109.593864] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: ece27044-3123-4685-a4ab-30eeb10337d1] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1109.601421] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.601743] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.601976] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.602197] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.602402] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.604780] env[61857]: INFO nova.compute.manager [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Terminating instance [ 1109.608430] env[61857]: INFO nova.scheduler.client.report [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleted allocations for instance 14df609d-678c-4713-8c9c-f80fabb61c79 [ 1109.609577] env[61857]: DEBUG nova.compute.manager [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1109.609869] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1109.610999] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be318c7e-5b50-4657-91e7-1166b230bc85 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.621554] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1109.621828] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f55504ba-e46f-4a13-b583-e441c54b5f3e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.631750] env[61857]: DEBUG oslo_vmware.api [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1109.631750] env[61857]: value = "task-951627" [ 1109.631750] env[61857]: _type = "Task" [ 1109.631750] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.639570] env[61857]: DEBUG oslo_vmware.api [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.097648] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 16b447b5-426a-4478-9d44-ae32b41dee50] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1110.120019] env[61857]: DEBUG oslo_concurrency.lockutils [None req-523e38b4-0556-4cee-93b1-d050524d1717 tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14df609d-678c-4713-8c9c-f80fabb61c79" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.485s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.139847] env[61857]: DEBUG oslo_vmware.api [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951627, 'name': PowerOffVM_Task, 'duration_secs': 0.19792} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.140146] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1110.140346] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1110.140600] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e854a138-f75b-4416-a0fe-b11f918f7a83 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.204502] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1110.204787] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1110.205019] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleting the datastore file [datastore2] 04438fc0-b5a1-4e29-84d5-4d84d332c8e4 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1110.205296] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4f138a7-0b7c-46cd-a675-937a42968c21 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.212479] env[61857]: DEBUG oslo_vmware.api [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1110.212479] env[61857]: value = "task-951629" [ 1110.212479] env[61857]: _type = "Task" [ 1110.212479] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.223280] env[61857]: DEBUG oslo_vmware.api [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951629, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.482497] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "14844876-96ab-481d-980d-4133e89d7d4d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.482497] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14844876-96ab-481d-980d-4133e89d7d4d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.482728] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "14844876-96ab-481d-980d-4133e89d7d4d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.482926] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14844876-96ab-481d-980d-4133e89d7d4d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.483119] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14844876-96ab-481d-980d-4133e89d7d4d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.485475] env[61857]: INFO nova.compute.manager [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Terminating instance [ 1110.487265] env[61857]: DEBUG nova.compute.manager [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1110.487468] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1110.488327] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6228a55f-6e42-4a35-bb3e-ab6b7c9cf485 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.495318] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquiring lock "857977a6-b81b-4c94-b823-436e57b34ba0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.495548] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lock "857977a6-b81b-4c94-b823-436e57b34ba0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.499240] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1110.499487] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4fdc0f05-4aa8-4974-a96a-7480c9d14ae1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.505728] env[61857]: DEBUG oslo_vmware.api [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1110.505728] env[61857]: value = "task-951630" [ 1110.505728] env[61857]: _type = "Task" [ 1110.505728] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.515151] env[61857]: DEBUG oslo_vmware.api [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.601231] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e5e64477-a1aa-4a4d-91a2-b17d912e09c3] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1110.678019] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2265a8c-8da6-404c-868a-f45ad3bbb865 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.688598] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67b1e58-3e5c-4ec8-b00e-10e04b16fb0c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.739719] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd7e9cc-5a46-47b9-ae17-885bece506e6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.750283] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c94d43-fff5-44f3-9f6b-960970cf14cf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.754134] env[61857]: DEBUG oslo_vmware.api [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951629, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245334} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.754408] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1110.754597] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1110.754777] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1110.754957] env[61857]: INFO nova.compute.manager [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1110.755222] env[61857]: DEBUG oslo.service.loopingcall [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.755814] env[61857]: DEBUG nova.compute.manager [-] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1110.755918] env[61857]: DEBUG nova.network.neutron [-] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1110.765553] env[61857]: DEBUG nova.compute.provider_tree [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1111.001017] env[61857]: DEBUG nova.compute.manager [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1111.018762] env[61857]: DEBUG oslo_vmware.api [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951630, 'name': PowerOffVM_Task, 'duration_secs': 0.220646} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.018762] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1111.018762] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1111.021390] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a0e8d08e-abe3-4c78-97e4-48a0e57b8062 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.048020] env[61857]: DEBUG oslo_vmware.rw_handles [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f77343-a6bd-6c0d-f45d-c0eef81c786c/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1111.048020] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49928c7-158a-46ba-b8b4-5b24838fbaa0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.053230] env[61857]: DEBUG oslo_vmware.rw_handles [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f77343-a6bd-6c0d-f45d-c0eef81c786c/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1111.053399] env[61857]: ERROR oslo_vmware.rw_handles [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f77343-a6bd-6c0d-f45d-c0eef81c786c/disk-0.vmdk due to incomplete transfer. [ 1111.053773] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-037c176e-b57a-408a-a99d-b41a804f7708 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.064128] env[61857]: DEBUG oslo_vmware.rw_handles [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f77343-a6bd-6c0d-f45d-c0eef81c786c/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1111.064128] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Uploaded image 788e8d57-21f0-4405-a84a-f5f5a69c06d1 to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1111.068415] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1111.068415] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-95f0c9f6-96d5-47b2-94e2-8a949203c1df {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.072965] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1111.072965] env[61857]: value = "task-951632" [ 1111.072965] env[61857]: _type = "Task" [ 1111.072965] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.081671] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951632, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.105456] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: a73f3163-b6fc-4764-b76a-451f823b5808] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1111.133832] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1111.134087] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1111.134274] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleting the datastore file [datastore1] 14844876-96ab-481d-980d-4133e89d7d4d {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1111.134579] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b030bf30-7937-4afb-8d98-b86d823478c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.142460] env[61857]: DEBUG oslo_vmware.api [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for the task: (returnval){ [ 1111.142460] env[61857]: value = "task-951633" [ 1111.142460] env[61857]: _type = "Task" [ 1111.142460] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.151523] env[61857]: DEBUG oslo_vmware.api [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951633, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.268517] env[61857]: DEBUG nova.scheduler.client.report [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1111.338425] env[61857]: DEBUG nova.compute.manager [req-e4a01e2a-8908-46ef-b326-6f906fb07b78 req-065a3a44-0158-4a6b-a028-00e24c2128a5 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Received event network-vif-deleted-83ff6e16-8b1b-4391-9ef1-57600502688c {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1111.338676] env[61857]: INFO nova.compute.manager [req-e4a01e2a-8908-46ef-b326-6f906fb07b78 req-065a3a44-0158-4a6b-a028-00e24c2128a5 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Neutron deleted interface 83ff6e16-8b1b-4391-9ef1-57600502688c; detaching it from the instance and deleting it from the info cache [ 1111.338906] env[61857]: DEBUG nova.network.neutron [req-e4a01e2a-8908-46ef-b326-6f906fb07b78 req-065a3a44-0158-4a6b-a028-00e24c2128a5 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.519747] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.583492] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951632, 'name': Destroy_Task, 'duration_secs': 0.326685} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.583771] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Destroyed the VM [ 1111.584013] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1111.584275] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3addf488-9c08-43c3-8dbe-739e4e45fcb5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.590486] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1111.590486] env[61857]: value = "task-951634" [ 1111.590486] env[61857]: _type = "Task" [ 1111.590486] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.599197] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951634, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.609249] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: bf2f46d4-5c6e-4493-b087-fe5a316b8c88] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1111.651778] env[61857]: DEBUG oslo_vmware.api [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Task: {'id': task-951633, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168043} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.652047] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1111.652241] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1111.652441] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1111.652627] env[61857]: INFO nova.compute.manager [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1111.652871] env[61857]: DEBUG oslo.service.loopingcall [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1111.653077] env[61857]: DEBUG nova.compute.manager [-] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1111.653175] env[61857]: DEBUG nova.network.neutron [-] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1111.773529] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.188s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.774065] env[61857]: DEBUG nova.compute.manager [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1111.776773] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.257s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.778747] env[61857]: INFO nova.compute.claims [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1111.814638] env[61857]: DEBUG nova.network.neutron [-] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.843410] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4f809ef-00a3-4448-a33c-6f75782d0a43 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.852053] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b99a2df-7f62-4680-bc6e-6ca6fc6d1208 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.877115] env[61857]: DEBUG nova.compute.manager [req-e4a01e2a-8908-46ef-b326-6f906fb07b78 req-065a3a44-0158-4a6b-a028-00e24c2128a5 service nova] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Detach interface failed, port_id=83ff6e16-8b1b-4391-9ef1-57600502688c, reason: Instance 04438fc0-b5a1-4e29-84d5-4d84d332c8e4 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1112.100856] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951634, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.112417] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 93e03e35-15a0-49e7-b1be-09178eabbfda] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1112.284289] env[61857]: DEBUG nova.compute.utils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1112.289016] env[61857]: DEBUG nova.compute.manager [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1112.289238] env[61857]: DEBUG nova.network.neutron [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1112.317830] env[61857]: INFO nova.compute.manager [-] [instance: 04438fc0-b5a1-4e29-84d5-4d84d332c8e4] Took 1.56 seconds to deallocate network for instance. [ 1112.364962] env[61857]: DEBUG nova.network.neutron [-] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.367686] env[61857]: DEBUG nova.policy [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98001d3024e44cb2a9cee667dac33574', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35ae3038a282412e963845e221e87f67', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1112.600479] env[61857]: DEBUG oslo_vmware.api [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951634, 'name': RemoveSnapshot_Task, 'duration_secs': 0.790604} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.600776] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1112.601150] env[61857]: INFO nova.compute.manager [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Took 14.04 seconds to snapshot the instance on the hypervisor. [ 1112.615566] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 96ff157c-4471-4a12-ad99-0aafd6c2dce5] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1112.678172] env[61857]: DEBUG nova.network.neutron [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Successfully created port: 9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1112.789962] env[61857]: DEBUG nova.compute.manager [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1112.824730] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.871430] env[61857]: INFO nova.compute.manager [-] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Took 1.22 seconds to deallocate network for instance. [ 1112.871995] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393af70c-fde8-4014-a802-1af2339d54e4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.884070] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01fb248c-f6e3-4918-ae9e-9e195156203c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.917253] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a5e6ed-3168-4002-876f-34ce320bb53a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.925508] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ff51e5-8072-45bb-83a9-96e0891cd8b3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.939224] env[61857]: DEBUG nova.compute.provider_tree [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.118573] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 6c4c1276-ae80-4f37-9ef9-4872dd415d24] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1113.147426] env[61857]: DEBUG nova.compute.manager [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Found 3 images (rotation: 2) {{(pid=61857) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1113.147645] env[61857]: DEBUG nova.compute.manager [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Rotating out 1 backups {{(pid=61857) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1113.147817] env[61857]: DEBUG nova.compute.manager [None req-67af3710-2ac3-47ac-a93e-ebc5ba2bcde6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Deleting image 3448d3d7-419d-4115-8a4c-1b26372f70aa {{(pid=61857) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1113.364639] env[61857]: DEBUG nova.compute.manager [req-b15e0513-33a9-4539-abb0-79b05061251d req-a476b66b-efea-4f49-b071-c3f026fb2ce0 service nova] [instance: 14844876-96ab-481d-980d-4133e89d7d4d] Received event network-vif-deleted-a5cec797-c084-4a3f-828a-2624cd347348 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1113.380832] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.442118] env[61857]: DEBUG nova.scheduler.client.report [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1113.622128] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 42d38f5e-628a-4030-85e1-3ec0595cf3c8] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1113.801764] env[61857]: DEBUG nova.compute.manager [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1113.821715] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1113.822060] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1113.822247] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.822718] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1113.822967] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.823158] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1113.823379] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1113.824608] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1113.824608] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1113.824608] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1113.824608] env[61857]: DEBUG nova.virt.hardware [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.825134] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8276a17c-146c-46d7-8196-d359f19e0239 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.833346] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e834d95-d2b3-4a9b-8b40-d81da7744521 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.947327] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.170s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.947946] env[61857]: DEBUG nova.compute.manager [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1113.950636] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.126s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.950843] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.952851] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.572s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.953083] env[61857]: DEBUG nova.objects.instance [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lazy-loading 'resources' on Instance uuid 14844876-96ab-481d-980d-4133e89d7d4d {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.975847] env[61857]: INFO nova.scheduler.client.report [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleted allocations for instance 04438fc0-b5a1-4e29-84d5-4d84d332c8e4 [ 1114.127194] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 24d3d515-e03a-4b4e-bb8e-bc18537125ac] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1114.216420] env[61857]: DEBUG nova.network.neutron [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Successfully updated port: 9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1114.456862] env[61857]: DEBUG nova.compute.utils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1114.461038] env[61857]: DEBUG nova.compute.manager [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1114.461424] env[61857]: DEBUG nova.network.neutron [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1114.483769] env[61857]: DEBUG oslo_concurrency.lockutils [None req-da834e1a-4174-4d8f-a3f2-00bba2d5b8bd tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "04438fc0-b5a1-4e29-84d5-4d84d332c8e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.882s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.509329] env[61857]: DEBUG nova.policy [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd743d096f51048a3bf9085fd8e4de81c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7ce5df26be54a4e8f258280a471f38b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1114.524010] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48383879-7a9e-4f57-a71b-dfc10584918a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.531995] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c17c40-27b8-46e9-a46c-88e025bedbc3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.564241] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d222e72c-8f4c-4ebf-b433-8132b8d14b89 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.572213] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46dad7e0-a2b2-4be8-b715-3b66e80ac32d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.585275] env[61857]: DEBUG nova.compute.provider_tree [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.628637] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 30d905ed-831d-44ae-807c-062de9a7e9fb] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1114.717138] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.718143] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.718143] env[61857]: DEBUG nova.network.neutron [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1114.848779] env[61857]: DEBUG nova.network.neutron [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Successfully created port: bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1114.961958] env[61857]: DEBUG nova.compute.manager [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1115.092379] env[61857]: DEBUG nova.scheduler.client.report [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1115.132812] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e91e99b9-1cd3-4345-af09-f14af4df1214] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1115.228521] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "857dbf02-d829-4476-9094-13a30e14c799" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.228766] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.228921] env[61857]: DEBUG nova.compute.manager [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1115.229829] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e83755-cfbb-470a-a584-d298adb8bb66 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.236439] env[61857]: DEBUG nova.compute.manager [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61857) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1115.237018] env[61857]: DEBUG nova.objects.instance [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'flavor' on Instance uuid 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.260746] env[61857]: DEBUG nova.network.neutron [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1115.390714] env[61857]: DEBUG nova.compute.manager [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received event network-vif-plugged-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1115.390943] env[61857]: DEBUG oslo_concurrency.lockutils [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] Acquiring lock "c6d20166-7251-406f-9003-b0599c2f5987-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.391171] env[61857]: DEBUG oslo_concurrency.lockutils [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] Lock "c6d20166-7251-406f-9003-b0599c2f5987-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.391349] env[61857]: DEBUG oslo_concurrency.lockutils [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] Lock "c6d20166-7251-406f-9003-b0599c2f5987-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.391523] env[61857]: DEBUG nova.compute.manager [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] No waiting events found dispatching network-vif-plugged-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1115.391691] env[61857]: WARNING nova.compute.manager [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received unexpected event network-vif-plugged-9a0b5327-2d9d-47ef-bb93-6900b47fae64 for instance with vm_state building and task_state spawning. [ 1115.391858] env[61857]: DEBUG nova.compute.manager [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received event network-changed-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1115.392026] env[61857]: DEBUG nova.compute.manager [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Refreshing instance network info cache due to event network-changed-9a0b5327-2d9d-47ef-bb93-6900b47fae64. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1115.392203] env[61857]: DEBUG oslo_concurrency.lockutils [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] Acquiring lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.419190] env[61857]: DEBUG nova.network.neutron [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updating instance_info_cache with network_info: [{"id": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "address": "fa:16:3e:91:72:71", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a0b5327-2d", "ovs_interfaceid": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.599942] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.647s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.624175] env[61857]: INFO nova.scheduler.client.report [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Deleted allocations for instance 14844876-96ab-481d-980d-4133e89d7d4d [ 1115.636414] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 36ca32ed-1ba2-40d0-93c4-086a26a3f9e5] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1115.688707] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "b33cfca7-5621-47a8-bb84-59d82847d699" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.688943] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "b33cfca7-5621-47a8-bb84-59d82847d699" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.741934] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1115.742192] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1433b5b-a7cb-4bce-a08d-c849903d952d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.748788] env[61857]: DEBUG oslo_vmware.api [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1115.748788] env[61857]: value = "task-951635" [ 1115.748788] env[61857]: _type = "Task" [ 1115.748788] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.757100] env[61857]: DEBUG oslo_vmware.api [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951635, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.921992] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.922321] env[61857]: DEBUG nova.compute.manager [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Instance network_info: |[{"id": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "address": "fa:16:3e:91:72:71", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a0b5327-2d", "ovs_interfaceid": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1115.922668] env[61857]: DEBUG oslo_concurrency.lockutils [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] Acquired lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.922909] env[61857]: DEBUG nova.network.neutron [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Refreshing network info cache for port 9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1115.924444] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:72:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a0b5327-2d9d-47ef-bb93-6900b47fae64', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1115.931590] env[61857]: DEBUG oslo.service.loopingcall [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.934527] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1115.935331] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c43e3168-4a5f-4b7e-8e9b-a52ae1bee695 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.954477] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1115.954477] env[61857]: value = "task-951636" [ 1115.954477] env[61857]: _type = "Task" [ 1115.954477] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.961727] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951636, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.971201] env[61857]: DEBUG nova.compute.manager [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1115.997312] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1115.997576] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1115.997736] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1115.997925] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1115.998090] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1115.998248] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1115.998457] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1115.998623] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1115.998802] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1115.998962] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1115.999156] env[61857]: DEBUG nova.virt.hardware [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1115.999990] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd171f06-267b-45c5-a178-78dae2dd2366 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.007256] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f0f923-33a9-40b8-b7df-0d9e669dd7e9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.132485] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a9c9102a-cf25-4608-8ccd-901e867b41ad tempest-AttachInterfacesTestJSON-1012022611 tempest-AttachInterfacesTestJSON-1012022611-project-member] Lock "14844876-96ab-481d-980d-4133e89d7d4d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.650s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.139301] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 28bf59ca-4ffe-4005-9a88-da0660ebb48a] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1116.190787] env[61857]: DEBUG nova.compute.manager [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1116.196135] env[61857]: DEBUG nova.network.neutron [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updated VIF entry in instance network info cache for port 9a0b5327-2d9d-47ef-bb93-6900b47fae64. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1116.196535] env[61857]: DEBUG nova.network.neutron [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updating instance_info_cache with network_info: [{"id": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "address": "fa:16:3e:91:72:71", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a0b5327-2d", "ovs_interfaceid": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.257919] env[61857]: DEBUG oslo_vmware.api [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951635, 'name': PowerOffVM_Task, 'duration_secs': 0.182172} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.259700] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1116.259700] env[61857]: DEBUG nova.compute.manager [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1116.259700] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381d7f9b-f450-4758-818e-908506ca3bca {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.351163] env[61857]: DEBUG nova.network.neutron [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Successfully updated port: bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1116.464250] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951636, 'name': CreateVM_Task, 'duration_secs': 0.29827} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.464457] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1116.465155] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.465330] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.465667] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1116.465923] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-666a25ab-2938-4648-8a94-64aec380aacc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.470407] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1116.470407] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b5aa1f-4052-c822-e1dc-258516e6eb5b" [ 1116.470407] env[61857]: _type = "Task" [ 1116.470407] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.477564] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b5aa1f-4052-c822-e1dc-258516e6eb5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.642667] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 71920a62-090d-4df1-937e-87df3b043e28] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1116.699654] env[61857]: DEBUG oslo_concurrency.lockutils [req-2c59c493-fa3e-47a4-b18f-9b30531f6c21 req-ecb46cdd-08ee-4ac0-b98c-af12bc6f48fa service nova] Releasing lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.712993] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.713131] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.714666] env[61857]: INFO nova.compute.claims [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1116.770583] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a4c7cdae-7224-499c-92c9-3272e43763fe tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.542s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.854195] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquiring lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.854451] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquired lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.854624] env[61857]: DEBUG nova.network.neutron [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1116.982886] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b5aa1f-4052-c822-e1dc-258516e6eb5b, 'name': SearchDatastore_Task, 'duration_secs': 0.011201} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.983330] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.983541] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1116.983834] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.983997] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.984209] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1116.984487] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f6e06569-0b1e-45cb-b9ed-934375e03857 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.993135] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1116.993327] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1116.994060] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddc87b0d-20eb-4c08-86e5-c25e825bf03b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.001158] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1117.001158] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad02e0-35ac-6ffd-2b7c-5f58505a0569" [ 1117.001158] env[61857]: _type = "Task" [ 1117.001158] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.007408] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad02e0-35ac-6ffd-2b7c-5f58505a0569, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.146065] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 781a2790-e317-46fe-9be8-40c9e5f1f771] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1117.406775] env[61857]: DEBUG nova.network.neutron [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1117.417996] env[61857]: DEBUG nova.compute.manager [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Received event network-vif-plugged-bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1117.418936] env[61857]: DEBUG oslo_concurrency.lockutils [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] Acquiring lock "857977a6-b81b-4c94-b823-436e57b34ba0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.419051] env[61857]: DEBUG oslo_concurrency.lockutils [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] Lock "857977a6-b81b-4c94-b823-436e57b34ba0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.419224] env[61857]: DEBUG oslo_concurrency.lockutils [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] Lock "857977a6-b81b-4c94-b823-436e57b34ba0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.419406] env[61857]: DEBUG nova.compute.manager [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] No waiting events found dispatching network-vif-plugged-bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1117.419582] env[61857]: WARNING nova.compute.manager [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Received unexpected event network-vif-plugged-bfc4be35-7059-4d0f-bb2f-99c368eb3840 for instance with vm_state building and task_state spawning. [ 1117.419748] env[61857]: DEBUG nova.compute.manager [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Received event network-changed-bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1117.419907] env[61857]: DEBUG nova.compute.manager [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Refreshing instance network info cache due to event network-changed-bfc4be35-7059-4d0f-bb2f-99c368eb3840. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1117.420094] env[61857]: DEBUG oslo_concurrency.lockutils [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] Acquiring lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.510065] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ad02e0-35ac-6ffd-2b7c-5f58505a0569, 'name': SearchDatastore_Task, 'duration_secs': 0.009262} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.510927] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afee2fd9-0288-4048-8c35-c02b911c8702 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.517159] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1117.517159] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fcfba6-eb00-e050-163b-14aead63d7d4" [ 1117.517159] env[61857]: _type = "Task" [ 1117.517159] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.529041] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fcfba6-eb00-e050-163b-14aead63d7d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.619286] env[61857]: DEBUG nova.network.neutron [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updating instance_info_cache with network_info: [{"id": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "address": "fa:16:3e:5d:01:97", "network": {"id": "dbaf4423-cac8-4e35-81f5-3a339a7a8299", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-321217147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f7ce5df26be54a4e8f258280a471f38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc4be35-70", "ovs_interfaceid": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.649139] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c4883f84-e62b-4b59-9484-270d82dc34e0] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1117.798896] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd32a59-be3e-4537-b14c-ae0f3d5164e4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.806987] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0bd139-5619-478b-a6e1-aaa92e5f3c4b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.836226] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69237f49-ea2b-4eb4-afe5-4aa648540cd6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.843291] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2b95b7-cf29-4895-824d-464243efeb51 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.855863] env[61857]: DEBUG nova.compute.provider_tree [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.028041] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fcfba6-eb00-e050-163b-14aead63d7d4, 'name': SearchDatastore_Task, 'duration_secs': 0.009507} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.028340] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.028595] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] c6d20166-7251-406f-9003-b0599c2f5987/c6d20166-7251-406f-9003-b0599c2f5987.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1118.028844] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e33357c4-8c61-4299-aef8-37f3a210c71d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.035588] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1118.035588] env[61857]: value = "task-951637" [ 1118.035588] env[61857]: _type = "Task" [ 1118.035588] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.042960] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.122488] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Releasing lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.122916] env[61857]: DEBUG nova.compute.manager [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Instance network_info: |[{"id": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "address": "fa:16:3e:5d:01:97", "network": {"id": "dbaf4423-cac8-4e35-81f5-3a339a7a8299", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-321217147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f7ce5df26be54a4e8f258280a471f38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc4be35-70", "ovs_interfaceid": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1118.123470] env[61857]: DEBUG oslo_concurrency.lockutils [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] Acquired lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.123564] env[61857]: DEBUG nova.network.neutron [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Refreshing network info cache for port bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1118.124911] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:01:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '02092ea4-bae0-4e42-b0ab-abc365b4395a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bfc4be35-7059-4d0f-bb2f-99c368eb3840', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1118.133183] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Creating folder: Project (f7ce5df26be54a4e8f258280a471f38b). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1118.137298] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e999ed65-390f-4f04-83be-81abe2407e2f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.147538] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Created folder: Project (f7ce5df26be54a4e8f258280a471f38b) in parent group-v214027. [ 1118.147818] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Creating folder: Instances. Parent ref: group-v214224. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1118.148015] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0d28e1e0-9e0f-4d45-abbc-be81fec9fe7b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.152175] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 7a800abe-ea0a-4077-a5da-dd60eab917e3] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1118.156872] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Created folder: Instances in parent group-v214224. [ 1118.157127] env[61857]: DEBUG oslo.service.loopingcall [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1118.157314] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1118.157512] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23a6d527-31bc-42fb-9ceb-fc65bb7e2804 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.176217] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1118.176217] env[61857]: value = "task-951640" [ 1118.176217] env[61857]: _type = "Task" [ 1118.176217] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.185707] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951640, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.359513] env[61857]: DEBUG nova.scheduler.client.report [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.459771] env[61857]: DEBUG nova.compute.manager [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Stashing vm_state: stopped {{(pid=61857) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1118.518162] env[61857]: DEBUG nova.network.neutron [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updated VIF entry in instance network info cache for port bfc4be35-7059-4d0f-bb2f-99c368eb3840. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1118.518363] env[61857]: DEBUG nova.network.neutron [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updating instance_info_cache with network_info: [{"id": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "address": "fa:16:3e:5d:01:97", "network": {"id": "dbaf4423-cac8-4e35-81f5-3a339a7a8299", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-321217147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f7ce5df26be54a4e8f258280a471f38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc4be35-70", "ovs_interfaceid": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.549365] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951637, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450429} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.549365] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] c6d20166-7251-406f-9003-b0599c2f5987/c6d20166-7251-406f-9003-b0599c2f5987.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1118.550382] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1118.550382] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-433afd16-1708-4217-9f88-2432ca41b8cc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.559101] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1118.559101] env[61857]: value = "task-951641" [ 1118.559101] env[61857]: _type = "Task" [ 1118.559101] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.567196] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951641, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.656089] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e286e6d0-25db-4d86-b9a2-3538bc3f624a] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1118.691569] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951640, 'name': CreateVM_Task, 'duration_secs': 0.360615} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.691856] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1118.692863] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.693189] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.693655] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1118.694048] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cea1db3e-14e9-46c6-a8e5-b8622edfc14b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.700117] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1118.700117] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52bda7f8-adab-1acd-3fae-e6b637de6030" [ 1118.700117] env[61857]: _type = "Task" [ 1118.700117] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.708355] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52bda7f8-adab-1acd-3fae-e6b637de6030, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.865951] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.153s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.866527] env[61857]: DEBUG nova.compute.manager [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1118.980904] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.981441] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.022165] env[61857]: DEBUG oslo_concurrency.lockutils [req-a82d17a4-23b4-4772-84b5-5c869da9ad9b req-9ba099df-8344-40de-b43c-07e73392c0ac service nova] Releasing lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.069933] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951641, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057101} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.070250] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1119.071044] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bba40dd-8309-43cb-aac8-d82fd0612fb0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.093409] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] c6d20166-7251-406f-9003-b0599c2f5987/c6d20166-7251-406f-9003-b0599c2f5987.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.093734] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f1ac6da-b4d1-44e4-82a2-b48778d01074 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.114260] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1119.114260] env[61857]: value = "task-951642" [ 1119.114260] env[61857]: _type = "Task" [ 1119.114260] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.122491] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951642, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.160614] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 3cacef7b-dd24-4d33-9500-bbff03e342bb] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1119.209799] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52bda7f8-adab-1acd-3fae-e6b637de6030, 'name': SearchDatastore_Task, 'duration_secs': 0.008054} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.210134] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.210414] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1119.210656] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.210809] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.210994] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1119.211272] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5438bcff-e9cd-4db5-9cbd-287762e547ea {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.219536] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1119.219725] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1119.220749] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a1a2d83-ac77-4143-8768-97f2508f3094 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.225825] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1119.225825] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52fb119b-17f5-fb65-e14b-4009b4f3ed86" [ 1119.225825] env[61857]: _type = "Task" [ 1119.225825] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.233513] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fb119b-17f5-fb65-e14b-4009b4f3ed86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.373020] env[61857]: DEBUG nova.compute.utils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1119.373020] env[61857]: DEBUG nova.compute.manager [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1119.373391] env[61857]: DEBUG nova.network.neutron [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1119.487029] env[61857]: INFO nova.compute.claims [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.583245] env[61857]: DEBUG nova.policy [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f719cbf1755d4c34bef92dab6679cb56', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ef111c8b3ff4f2383b4e81db026792b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1119.624575] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951642, 'name': ReconfigVM_Task, 'duration_secs': 0.438464} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.624889] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Reconfigured VM instance instance-0000006a to attach disk [datastore2] c6d20166-7251-406f-9003-b0599c2f5987/c6d20166-7251-406f-9003-b0599c2f5987.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.625520] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be3f9b4a-4b6a-4208-a0ba-1d423418be1a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.631919] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1119.631919] env[61857]: value = "task-951643" [ 1119.631919] env[61857]: _type = "Task" [ 1119.631919] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.639412] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951643, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.663706] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: d49008e6-7789-4785-8eea-9b3fa36c4ccb] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1119.736054] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52fb119b-17f5-fb65-e14b-4009b4f3ed86, 'name': SearchDatastore_Task, 'duration_secs': 0.011104} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.736937] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bbe423e-644f-431a-81c6-0b01bc57381b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.743782] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1119.743782] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]522cb6c1-11c8-85fd-d3b7-4032d3c999e8" [ 1119.743782] env[61857]: _type = "Task" [ 1119.743782] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.751229] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522cb6c1-11c8-85fd-d3b7-4032d3c999e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.878133] env[61857]: DEBUG nova.compute.manager [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1119.952114] env[61857]: DEBUG nova.network.neutron [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Successfully created port: 1c8b2449-1a4b-4062-a760-d88145f9615e {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1119.994349] env[61857]: INFO nova.compute.resource_tracker [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating resource usage from migration 955e4413-8c36-40ff-8439-c037905e8eef [ 1120.076478] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afe2d4d-f5da-488b-be7e-01f846035632 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.084404] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fab76c-849a-49ad-ac69-836a3e57c4c9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.115520] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9dc27b-5148-4660-83c8-056323d4128b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.122823] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec73d7ba-3aea-46d0-a2fe-51cf514528dc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.136132] env[61857]: DEBUG nova.compute.provider_tree [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.146309] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951643, 'name': Rename_Task, 'duration_secs': 0.268391} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.147115] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1120.147536] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ea8530a-3704-4325-8cd9-3ca89a5cbd99 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.154171] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1120.154171] env[61857]: value = "task-951644" [ 1120.154171] env[61857]: _type = "Task" [ 1120.154171] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.166341] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951644, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.166736] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 18d9307d-6a20-47b4-b4e9-176f9a8c33cc] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1120.253419] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522cb6c1-11c8-85fd-d3b7-4032d3c999e8, 'name': SearchDatastore_Task, 'duration_secs': 0.011963} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.253697] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.253964] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 857977a6-b81b-4c94-b823-436e57b34ba0/857977a6-b81b-4c94-b823-436e57b34ba0.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1120.254286] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32ccddf5-55dc-4261-8523-1723317c653b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.260997] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1120.260997] env[61857]: value = "task-951645" [ 1120.260997] env[61857]: _type = "Task" [ 1120.260997] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.268327] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951645, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.643151] env[61857]: DEBUG nova.scheduler.client.report [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1120.665842] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951644, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.670764] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c4f70321-f612-416e-b209-eb974dab9d49] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1120.770657] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951645, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460502} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.770938] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 857977a6-b81b-4c94-b823-436e57b34ba0/857977a6-b81b-4c94-b823-436e57b34ba0.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1120.771177] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1120.771772] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-221ee844-7fad-4270-bced-f1fd344450c5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.778226] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1120.778226] env[61857]: value = "task-951646" [ 1120.778226] env[61857]: _type = "Task" [ 1120.778226] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.785523] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951646, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.889836] env[61857]: DEBUG nova.compute.manager [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1120.917579] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1120.917844] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1120.918014] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1120.918231] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1120.918428] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1120.918592] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1120.918803] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1120.918967] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1120.919161] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1120.919330] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1120.919509] env[61857]: DEBUG nova.virt.hardware [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1120.920413] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abce6761-b1c8-4b25-bddb-d9dfb16351f9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.928610] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7418c34c-7c6b-4424-8a7d-cf9e48af9426 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.148863] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.167s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.149126] env[61857]: INFO nova.compute.manager [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Migrating [ 1121.173935] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: e88aed62-6561-4e29-b041-46b19cc5ef63] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1121.175844] env[61857]: DEBUG oslo_vmware.api [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951644, 'name': PowerOnVM_Task, 'duration_secs': 0.602187} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.176546] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1121.176758] env[61857]: INFO nova.compute.manager [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Took 7.37 seconds to spawn the instance on the hypervisor. [ 1121.177368] env[61857]: DEBUG nova.compute.manager [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1121.178476] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a926954-ccd1-4520-be8c-cf9f4108c7b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.296061] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951646, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05822} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.296419] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1121.297462] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90fcb69-88b8-4074-83d8-20bffce27733 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.323467] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 857977a6-b81b-4c94-b823-436e57b34ba0/857977a6-b81b-4c94-b823-436e57b34ba0.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1121.323798] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e523011-0525-42ef-85b4-b7de24688669 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.350963] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1121.350963] env[61857]: value = "task-951647" [ 1121.350963] env[61857]: _type = "Task" [ 1121.350963] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.362713] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951647, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.370612] env[61857]: DEBUG nova.compute.manager [req-fde11a78-3b63-49ed-8a26-9b95c41d20e6 req-82bd5998-939f-4218-82a9-c059f070bb10 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Received event network-vif-plugged-1c8b2449-1a4b-4062-a760-d88145f9615e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1121.370974] env[61857]: DEBUG oslo_concurrency.lockutils [req-fde11a78-3b63-49ed-8a26-9b95c41d20e6 req-82bd5998-939f-4218-82a9-c059f070bb10 service nova] Acquiring lock "b33cfca7-5621-47a8-bb84-59d82847d699-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.371305] env[61857]: DEBUG oslo_concurrency.lockutils [req-fde11a78-3b63-49ed-8a26-9b95c41d20e6 req-82bd5998-939f-4218-82a9-c059f070bb10 service nova] Lock "b33cfca7-5621-47a8-bb84-59d82847d699-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.371827] env[61857]: DEBUG oslo_concurrency.lockutils [req-fde11a78-3b63-49ed-8a26-9b95c41d20e6 req-82bd5998-939f-4218-82a9-c059f070bb10 service nova] Lock "b33cfca7-5621-47a8-bb84-59d82847d699-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.372101] env[61857]: DEBUG nova.compute.manager [req-fde11a78-3b63-49ed-8a26-9b95c41d20e6 req-82bd5998-939f-4218-82a9-c059f070bb10 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] No waiting events found dispatching network-vif-plugged-1c8b2449-1a4b-4062-a760-d88145f9615e {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1121.372603] env[61857]: WARNING nova.compute.manager [req-fde11a78-3b63-49ed-8a26-9b95c41d20e6 req-82bd5998-939f-4218-82a9-c059f070bb10 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Received unexpected event network-vif-plugged-1c8b2449-1a4b-4062-a760-d88145f9615e for instance with vm_state building and task_state spawning. [ 1121.456393] env[61857]: DEBUG nova.network.neutron [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Successfully updated port: 1c8b2449-1a4b-4062-a760-d88145f9615e {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1121.671250] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.672010] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.672010] env[61857]: DEBUG nova.network.neutron [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1121.677110] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 11adaee5-8e56-4679-8373-5d5690a44ca9] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1121.702618] env[61857]: INFO nova.compute.manager [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Took 13.11 seconds to build instance. [ 1121.862176] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951647, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.960226] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.960437] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.960610] env[61857]: DEBUG nova.network.neutron [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1122.180177] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 2c187f64-8a36-4dfd-94e3-8ea944dbac24] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1122.204733] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9db87c45-3df2-42fe-a720-c36bfb183138 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "c6d20166-7251-406f-9003-b0599c2f5987" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.629s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.366356] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951647, 'name': ReconfigVM_Task, 'duration_secs': 0.609722} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.366912] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 857977a6-b81b-4c94-b823-436e57b34ba0/857977a6-b81b-4c94-b823-436e57b34ba0.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1122.367627] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c51b4ee-6430-4a52-82fe-ea031d787563 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.375320] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1122.375320] env[61857]: value = "task-951648" [ 1122.375320] env[61857]: _type = "Task" [ 1122.375320] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.387284] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951648, 'name': Rename_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.502313] env[61857]: DEBUG nova.network.neutron [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1122.660177] env[61857]: DEBUG nova.network.neutron [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Updating instance_info_cache with network_info: [{"id": "1c8b2449-1a4b-4062-a760-d88145f9615e", "address": "fa:16:3e:b0:fd:68", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c8b2449-1a", "ovs_interfaceid": "1c8b2449-1a4b-4062-a760-d88145f9615e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.679137] env[61857]: DEBUG nova.network.neutron [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance_info_cache with network_info: [{"id": "f3163567-eee7-4192-9332-fe1748882d93", "address": "fa:16:3e:80:c4:35", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3163567-ee", "ovs_interfaceid": "f3163567-eee7-4192-9332-fe1748882d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.888592] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951648, 'name': Rename_Task, 'duration_secs': 0.133923} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.889062] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1122.889433] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-445220ca-572c-475c-9c4b-c9cdeb668f3b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.895288] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1122.895288] env[61857]: value = "task-951649" [ 1122.895288] env[61857]: _type = "Task" [ 1122.895288] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.904468] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951649, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.163431] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.163537] env[61857]: DEBUG nova.compute.manager [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Instance network_info: |[{"id": "1c8b2449-1a4b-4062-a760-d88145f9615e", "address": "fa:16:3e:b0:fd:68", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c8b2449-1a", "ovs_interfaceid": "1c8b2449-1a4b-4062-a760-d88145f9615e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1123.164030] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:fd:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c8b2449-1a4b-4062-a760-d88145f9615e', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1123.173160] env[61857]: DEBUG oslo.service.loopingcall [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1123.173447] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1123.173701] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08a18e6c-baff-4a9c-b89a-eef208e8afb6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.189162] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.199050] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1123.199050] env[61857]: value = "task-951650" [ 1123.199050] env[61857]: _type = "Task" [ 1123.199050] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.209802] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951650, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.408785] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951649, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.414279] env[61857]: DEBUG nova.compute.manager [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Received event network-changed-1c8b2449-1a4b-4062-a760-d88145f9615e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1123.414279] env[61857]: DEBUG nova.compute.manager [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Refreshing instance network info cache due to event network-changed-1c8b2449-1a4b-4062-a760-d88145f9615e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1123.414887] env[61857]: DEBUG oslo_concurrency.lockutils [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] Acquiring lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.415100] env[61857]: DEBUG oslo_concurrency.lockutils [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] Acquired lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.415336] env[61857]: DEBUG nova.network.neutron [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Refreshing network info cache for port 1c8b2449-1a4b-4062-a760-d88145f9615e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1123.492390] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquiring lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.492590] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.708453] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951650, 'name': CreateVM_Task, 'duration_secs': 0.360524} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.708863] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1123.709341] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.709608] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.709876] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1123.710147] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c13cdf2a-8216-4e1e-9b6b-ffdeae5abe86 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.714634] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1123.714634] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523ab926-e46c-5b26-1e72-de08021f9576" [ 1123.714634] env[61857]: _type = "Task" [ 1123.714634] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.726195] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523ab926-e46c-5b26-1e72-de08021f9576, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.894873] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.907075] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951649, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.994799] env[61857]: DEBUG nova.compute.manager [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1124.167126] env[61857]: DEBUG nova.network.neutron [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Updated VIF entry in instance network info cache for port 1c8b2449-1a4b-4062-a760-d88145f9615e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1124.167523] env[61857]: DEBUG nova.network.neutron [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Updating instance_info_cache with network_info: [{"id": "1c8b2449-1a4b-4062-a760-d88145f9615e", "address": "fa:16:3e:b0:fd:68", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c8b2449-1a", "ovs_interfaceid": "1c8b2449-1a4b-4062-a760-d88145f9615e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.224578] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523ab926-e46c-5b26-1e72-de08021f9576, 'name': SearchDatastore_Task, 'duration_secs': 0.013478} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.224945] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.225251] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1124.225502] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.225661] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.225845] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1124.226119] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b4278c5-c66e-476a-9d91-394a9994ae47 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.234429] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1124.234617] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1124.235413] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1f14cfb-6c52-4f46-9152-1842fa161438 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.242769] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1124.242769] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52f570f4-0948-09fb-5ce8-64efd90a4b9a" [ 1124.242769] env[61857]: _type = "Task" [ 1124.242769] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.250800] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f570f4-0948-09fb-5ce8-64efd90a4b9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.398932] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Getting list of instances from cluster (obj){ [ 1124.398932] env[61857]: value = "domain-c8" [ 1124.398932] env[61857]: _type = "ClusterComputeResource" [ 1124.398932] env[61857]: } {{(pid=61857) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1124.400143] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-191f1d77-b37e-4e60-8a5b-579c78338f51 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.418699] env[61857]: DEBUG oslo_vmware.api [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951649, 'name': PowerOnVM_Task, 'duration_secs': 1.458892} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.418961] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Got total of 4 instances {{(pid=61857) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1124.419140] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Triggering sync for uuid 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1124.419341] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Triggering sync for uuid c6d20166-7251-406f-9003-b0599c2f5987 {{(pid=61857) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1124.419518] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Triggering sync for uuid 857977a6-b81b-4c94-b823-436e57b34ba0 {{(pid=61857) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1124.419640] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Triggering sync for uuid b33cfca7-5621-47a8-bb84-59d82847d699 {{(pid=61857) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1124.419951] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1124.420385] env[61857]: INFO nova.compute.manager [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Took 8.45 seconds to spawn the instance on the hypervisor. [ 1124.420385] env[61857]: DEBUG nova.compute.manager [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1124.420680] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "857dbf02-d829-4476-9094-13a30e14c799" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.420899] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "857dbf02-d829-4476-9094-13a30e14c799" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.421083] env[61857]: INFO nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 857dbf02-d829-4476-9094-13a30e14c799] During sync_power_state the instance has a pending task (resize_migrating). Skip. [ 1124.421261] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "857dbf02-d829-4476-9094-13a30e14c799" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.421492] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "c6d20166-7251-406f-9003-b0599c2f5987" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.421685] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "c6d20166-7251-406f-9003-b0599c2f5987" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.421923] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "857977a6-b81b-4c94-b823-436e57b34ba0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.422777] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "b33cfca7-5621-47a8-bb84-59d82847d699" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.422974] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe64f429-aa56-4d9f-b8db-b51799869a1b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.425833] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80bf127-62fd-4826-91ff-d71d0a50bd90 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.515816] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.516101] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.517543] env[61857]: INFO nova.compute.claims [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.670399] env[61857]: DEBUG oslo_concurrency.lockutils [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] Releasing lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.670695] env[61857]: DEBUG nova.compute.manager [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received event network-changed-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1124.670875] env[61857]: DEBUG nova.compute.manager [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Refreshing instance network info cache due to event network-changed-9a0b5327-2d9d-47ef-bb93-6900b47fae64. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1124.671115] env[61857]: DEBUG oslo_concurrency.lockutils [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] Acquiring lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.671266] env[61857]: DEBUG oslo_concurrency.lockutils [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] Acquired lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.671453] env[61857]: DEBUG nova.network.neutron [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Refreshing network info cache for port 9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1124.705314] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fe09e2-cc90-4ca0-a177-0776bf878613 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.723390] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance '857dbf02-d829-4476-9094-13a30e14c799' progress to 0 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1124.751988] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f570f4-0948-09fb-5ce8-64efd90a4b9a, 'name': SearchDatastore_Task, 'duration_secs': 0.009404} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.752741] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-051d9a15-36ee-4121-9793-1081a3e48a8f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.757638] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1124.757638] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524da72e-9d91-1c6e-3336-e832437f937f" [ 1124.757638] env[61857]: _type = "Task" [ 1124.757638] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.765304] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524da72e-9d91-1c6e-3336-e832437f937f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.941122] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "c6d20166-7251-406f-9003-b0599c2f5987" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.519s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.944842] env[61857]: INFO nova.compute.manager [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Took 13.44 seconds to build instance. [ 1125.229227] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1125.229546] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2785f52a-cd70-4d52-b674-ce2c07c240c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.237094] env[61857]: DEBUG oslo_vmware.api [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1125.237094] env[61857]: value = "task-951651" [ 1125.237094] env[61857]: _type = "Task" [ 1125.237094] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.247634] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] VM already powered off {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1125.247857] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance '857dbf02-d829-4476-9094-13a30e14c799' progress to 17 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1125.266177] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524da72e-9d91-1c6e-3336-e832437f937f, 'name': SearchDatastore_Task, 'duration_secs': 0.009421} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.266515] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.266739] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] b33cfca7-5621-47a8-bb84-59d82847d699/b33cfca7-5621-47a8-bb84-59d82847d699.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1125.266995] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d481a550-4c84-4fa3-9fc7-f52c6c443af5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.273016] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1125.273016] env[61857]: value = "task-951652" [ 1125.273016] env[61857]: _type = "Task" [ 1125.273016] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.280739] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.436968] env[61857]: DEBUG nova.network.neutron [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updated VIF entry in instance network info cache for port 9a0b5327-2d9d-47ef-bb93-6900b47fae64. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1125.437466] env[61857]: DEBUG nova.network.neutron [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updating instance_info_cache with network_info: [{"id": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "address": "fa:16:3e:91:72:71", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a0b5327-2d", "ovs_interfaceid": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.446679] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ed83658-c48e-4e30-bb34-1f95ebf42815 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lock "857977a6-b81b-4c94-b823-436e57b34ba0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.951s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.446997] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "857977a6-b81b-4c94-b823-436e57b34ba0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.025s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.447218] env[61857]: INFO nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] During sync_power_state the instance has a pending task (spawning). Skip. [ 1125.447798] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "857977a6-b81b-4c94-b823-436e57b34ba0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.624744] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0929d867-55e2-4788-80b8-41d4684ef2ae {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.633233] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fc8353-6e38-4b23-bdbd-cc02ab9bc969 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.665104] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772db4ff-f496-42eb-a47c-2076e808483e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.673215] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16049dd4-9d59-452b-ae99-03736414b326 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.689036] env[61857]: DEBUG nova.compute.provider_tree [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.754746] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1125.755018] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1125.755191] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1125.755531] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1125.755531] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1125.755693] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1125.755900] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1125.756078] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1125.756248] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1125.756454] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1125.756771] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1125.761908] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b23efb0-0554-4288-b9e7-828ec870539a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.778629] env[61857]: DEBUG oslo_vmware.api [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1125.778629] env[61857]: value = "task-951653" [ 1125.778629] env[61857]: _type = "Task" [ 1125.778629] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.784934] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951652, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437382} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.785488] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] b33cfca7-5621-47a8-bb84-59d82847d699/b33cfca7-5621-47a8-bb84-59d82847d699.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1125.785710] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1125.785944] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f7183b1-6a18-4253-8f29-66ba8ad681ad {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.790609] env[61857]: DEBUG oslo_vmware.api [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951653, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.795168] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1125.795168] env[61857]: value = "task-951654" [ 1125.795168] env[61857]: _type = "Task" [ 1125.795168] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.802784] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951654, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.940136] env[61857]: DEBUG oslo_concurrency.lockutils [req-59060f90-ce4c-4742-9063-8396b2cdeb79 req-558e97cd-573c-414d-a044-5928cb9b3040 service nova] Releasing lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.192216] env[61857]: DEBUG nova.scheduler.client.report [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.237847] env[61857]: INFO nova.compute.manager [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Rescuing [ 1126.238207] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquiring lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.238314] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquired lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.238518] env[61857]: DEBUG nova.network.neutron [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1126.288858] env[61857]: DEBUG oslo_vmware.api [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951653, 'name': ReconfigVM_Task, 'duration_secs': 0.151115} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.289157] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance '857dbf02-d829-4476-9094-13a30e14c799' progress to 33 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1126.303874] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951654, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059695} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.304151] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1126.304894] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74dfd11b-2849-4bb1-8944-c96579fd951d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.327863] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] b33cfca7-5621-47a8-bb84-59d82847d699/b33cfca7-5621-47a8-bb84-59d82847d699.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1126.328167] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30c50e2f-2570-47c7-abc7-b5febf403869 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.347951] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1126.347951] env[61857]: value = "task-951655" [ 1126.347951] env[61857]: _type = "Task" [ 1126.347951] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.355831] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951655, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.697065] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.795633] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1126.795888] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1126.796070] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1126.796270] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1126.796427] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1126.796584] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1126.796791] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1126.796956] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1126.797144] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1126.797316] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1126.797498] env[61857]: DEBUG nova.virt.hardware [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1126.802847] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Reconfiguring VM instance instance-00000061 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1126.803134] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-029b09e8-4d69-4851-89f3-eefa433e8324 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.824366] env[61857]: DEBUG oslo_vmware.api [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1126.824366] env[61857]: value = "task-951656" [ 1126.824366] env[61857]: _type = "Task" [ 1126.824366] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.833713] env[61857]: DEBUG oslo_vmware.api [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951656, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.856632] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951655, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.982066] env[61857]: DEBUG nova.network.neutron [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updating instance_info_cache with network_info: [{"id": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "address": "fa:16:3e:5d:01:97", "network": {"id": "dbaf4423-cac8-4e35-81f5-3a339a7a8299", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-321217147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f7ce5df26be54a4e8f258280a471f38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc4be35-70", "ovs_interfaceid": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.200310] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquiring lock "cec40198-bb15-435c-ae8b-cdf98dad3574" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.200662] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "cec40198-bb15-435c-ae8b-cdf98dad3574" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.334742] env[61857]: DEBUG oslo_vmware.api [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951656, 'name': ReconfigVM_Task, 'duration_secs': 0.3536} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.335126] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Reconfigured VM instance instance-00000061 to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1127.335927] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645daff4-a202-4862-901b-8d3b7e1d534b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.358102] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 857dbf02-d829-4476-9094-13a30e14c799/857dbf02-d829-4476-9094-13a30e14c799.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.361156] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-791fc778-c685-4f5c-a938-f500824b3642 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.379548] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951655, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.380917] env[61857]: DEBUG oslo_vmware.api [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1127.380917] env[61857]: value = "task-951657" [ 1127.380917] env[61857]: _type = "Task" [ 1127.380917] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.388613] env[61857]: DEBUG oslo_vmware.api [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951657, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.485077] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Releasing lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.704073] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "cec40198-bb15-435c-ae8b-cdf98dad3574" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.503s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.704528] env[61857]: DEBUG nova.compute.manager [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1127.860133] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951655, 'name': ReconfigVM_Task, 'duration_secs': 1.253879} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.860449] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Reconfigured VM instance instance-0000006c to attach disk [datastore1] b33cfca7-5621-47a8-bb84-59d82847d699/b33cfca7-5621-47a8-bb84-59d82847d699.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1127.861079] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0190b126-3f3e-4e40-8273-c5536ba42c7a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.868895] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1127.868895] env[61857]: value = "task-951658" [ 1127.868895] env[61857]: _type = "Task" [ 1127.868895] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.876343] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951658, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.888456] env[61857]: DEBUG oslo_vmware.api [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951657, 'name': ReconfigVM_Task, 'duration_secs': 0.258925} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.888706] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 857dbf02-d829-4476-9094-13a30e14c799/857dbf02-d829-4476-9094-13a30e14c799.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1127.888974] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance '857dbf02-d829-4476-9094-13a30e14c799' progress to 50 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1128.019339] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1128.019489] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76be834d-f466-4a75-8e3d-c2db52f51023 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.026030] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1128.026030] env[61857]: value = "task-951659" [ 1128.026030] env[61857]: _type = "Task" [ 1128.026030] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.036502] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951659, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.209366] env[61857]: DEBUG nova.compute.utils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1128.210917] env[61857]: DEBUG nova.compute.manager [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1128.211110] env[61857]: DEBUG nova.network.neutron [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1128.249358] env[61857]: DEBUG nova.policy [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f18cda0a217843588476e8ce34c1e2b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38a031bc6d084589a12d2246c95c0324', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1128.379522] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951658, 'name': Rename_Task, 'duration_secs': 0.12015} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.379869] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1128.380038] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cf2362a-f4b4-49ea-b2d2-1ea56186cd83 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.386999] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1128.386999] env[61857]: value = "task-951660" [ 1128.386999] env[61857]: _type = "Task" [ 1128.386999] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.396085] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951660, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.396797] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6386a3dd-3452-4df4-95d6-7d1380821483 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.415617] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2353024-85f0-4d13-a0af-731afb2b21a1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.433614] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance '857dbf02-d829-4476-9094-13a30e14c799' progress to 67 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1128.519305] env[61857]: DEBUG nova.network.neutron [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Successfully created port: 9d257050-d1bb-4736-9db7-98fe7964a3e4 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1128.535365] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951659, 'name': PowerOffVM_Task, 'duration_secs': 0.143249} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.535742] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1128.536848] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70897cc6-6ef9-4ace-9127-f9a8f8001155 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.555708] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6b99fc-643d-4f22-bc18-4a8bbc7afa7d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.583404] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1128.583746] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0da19d83-eed1-4cbd-be9c-b9084d2a6ed9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.592332] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1128.592332] env[61857]: value = "task-951661" [ 1128.592332] env[61857]: _type = "Task" [ 1128.592332] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.601770] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] VM already powered off {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1128.602010] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1128.602263] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.602420] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.602608] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.602866] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8dfffd1b-2299-414b-b111-360d70b01b29 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.613413] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.613806] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1128.614569] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee012323-af0a-429f-b485-cfdc41a398b9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.621579] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1128.621579] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524b0c28-6068-ac0a-adef-113841c7ec3d" [ 1128.621579] env[61857]: _type = "Task" [ 1128.621579] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.628850] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524b0c28-6068-ac0a-adef-113841c7ec3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.714032] env[61857]: DEBUG nova.compute.manager [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1128.896535] env[61857]: DEBUG oslo_vmware.api [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951660, 'name': PowerOnVM_Task, 'duration_secs': 0.469776} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.896815] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1128.897030] env[61857]: INFO nova.compute.manager [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Took 8.01 seconds to spawn the instance on the hypervisor. [ 1128.897222] env[61857]: DEBUG nova.compute.manager [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1128.898314] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c358a4-a90b-4e2d-8cbe-4ea6fccaef9d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.976985] env[61857]: DEBUG nova.network.neutron [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Port f3163567-eee7-4192-9332-fe1748882d93 binding to destination host cpu-1 is already ACTIVE {{(pid=61857) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1129.133030] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]524b0c28-6068-ac0a-adef-113841c7ec3d, 'name': SearchDatastore_Task, 'duration_secs': 0.011652} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.133841] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ecb8e82-918c-4d77-a4a7-dbe1871c08b7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.138778] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1129.138778] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]529e0b39-ce1b-076c-f9d4-44022d061e6f" [ 1129.138778] env[61857]: _type = "Task" [ 1129.138778] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.146210] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529e0b39-ce1b-076c-f9d4-44022d061e6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.415098] env[61857]: INFO nova.compute.manager [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Took 12.72 seconds to build instance. [ 1129.649167] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]529e0b39-ce1b-076c-f9d4-44022d061e6f, 'name': SearchDatastore_Task, 'duration_secs': 0.008883} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.649531] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.649835] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 857977a6-b81b-4c94-b823-436e57b34ba0/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk. {{(pid=61857) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1129.650148] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-562de102-e53d-49cf-9fc6-724cde7aff1f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.656855] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1129.656855] env[61857]: value = "task-951662" [ 1129.656855] env[61857]: _type = "Task" [ 1129.656855] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.664115] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951662, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.723939] env[61857]: DEBUG nova.compute.manager [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1129.749262] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1129.749525] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1129.749707] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1129.749901] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1129.750068] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1129.750228] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1129.750571] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1129.750762] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1129.751039] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1129.751330] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1129.751557] env[61857]: DEBUG nova.virt.hardware [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1129.752428] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd975cc1-6068-4677-9856-32407a642273 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.759968] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c51c3d-62df-4be4-a3b1-0149ca3b24b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.916988] env[61857]: DEBUG oslo_concurrency.lockutils [None req-647a6275-b9f8-4ebf-95c1-4ac5663e85ad tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "b33cfca7-5621-47a8-bb84-59d82847d699" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.228s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.917340] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "b33cfca7-5621-47a8-bb84-59d82847d699" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.495s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.917547] env[61857]: INFO nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] During sync_power_state the instance has a pending task (spawning). Skip. [ 1129.917739] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "b33cfca7-5621-47a8-bb84-59d82847d699" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.999594] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "857dbf02-d829-4476-9094-13a30e14c799-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.000024] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.000284] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.050570] env[61857]: DEBUG nova.compute.manager [req-8cd07df4-077e-433c-952e-d0d055594805 req-1eeffd79-5bd3-4b97-a46d-5b79a7b4f379 service nova] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Received event network-vif-plugged-9d257050-d1bb-4736-9db7-98fe7964a3e4 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1130.051045] env[61857]: DEBUG oslo_concurrency.lockutils [req-8cd07df4-077e-433c-952e-d0d055594805 req-1eeffd79-5bd3-4b97-a46d-5b79a7b4f379 service nova] Acquiring lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.051469] env[61857]: DEBUG oslo_concurrency.lockutils [req-8cd07df4-077e-433c-952e-d0d055594805 req-1eeffd79-5bd3-4b97-a46d-5b79a7b4f379 service nova] Lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.051831] env[61857]: DEBUG oslo_concurrency.lockutils [req-8cd07df4-077e-433c-952e-d0d055594805 req-1eeffd79-5bd3-4b97-a46d-5b79a7b4f379 service nova] Lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.052189] env[61857]: DEBUG nova.compute.manager [req-8cd07df4-077e-433c-952e-d0d055594805 req-1eeffd79-5bd3-4b97-a46d-5b79a7b4f379 service nova] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] No waiting events found dispatching network-vif-plugged-9d257050-d1bb-4736-9db7-98fe7964a3e4 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1130.052527] env[61857]: WARNING nova.compute.manager [req-8cd07df4-077e-433c-952e-d0d055594805 req-1eeffd79-5bd3-4b97-a46d-5b79a7b4f379 service nova] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Received unexpected event network-vif-plugged-9d257050-d1bb-4736-9db7-98fe7964a3e4 for instance with vm_state building and task_state spawning. [ 1130.147863] env[61857]: DEBUG nova.network.neutron [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Successfully updated port: 9d257050-d1bb-4736-9db7-98fe7964a3e4 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1130.170615] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951662, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43934} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.172287] env[61857]: INFO nova.virt.vmwareapi.ds_util [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 857977a6-b81b-4c94-b823-436e57b34ba0/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk. [ 1130.172287] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279daab3-5c7d-44d3-9201-85da19e32bc8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.200268] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 857977a6-b81b-4c94-b823-436e57b34ba0/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1130.200896] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a79c350-6cbf-47a6-92d3-f15d5a72ac46 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.219777] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1130.219777] env[61857]: value = "task-951663" [ 1130.219777] env[61857]: _type = "Task" [ 1130.219777] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.229300] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951663, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.506732] env[61857]: DEBUG nova.compute.manager [req-dd5112cf-86b5-426b-a5ac-c966fb01b164 req-9c48766f-9f51-46a6-b30a-0a31054087a0 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Received event network-changed-1c8b2449-1a4b-4062-a760-d88145f9615e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1130.506969] env[61857]: DEBUG nova.compute.manager [req-dd5112cf-86b5-426b-a5ac-c966fb01b164 req-9c48766f-9f51-46a6-b30a-0a31054087a0 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Refreshing instance network info cache due to event network-changed-1c8b2449-1a4b-4062-a760-d88145f9615e. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1130.507172] env[61857]: DEBUG oslo_concurrency.lockutils [req-dd5112cf-86b5-426b-a5ac-c966fb01b164 req-9c48766f-9f51-46a6-b30a-0a31054087a0 service nova] Acquiring lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.507277] env[61857]: DEBUG oslo_concurrency.lockutils [req-dd5112cf-86b5-426b-a5ac-c966fb01b164 req-9c48766f-9f51-46a6-b30a-0a31054087a0 service nova] Acquired lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.507442] env[61857]: DEBUG nova.network.neutron [req-dd5112cf-86b5-426b-a5ac-c966fb01b164 req-9c48766f-9f51-46a6-b30a-0a31054087a0 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Refreshing network info cache for port 1c8b2449-1a4b-4062-a760-d88145f9615e {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1130.656474] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquiring lock "refresh_cache-fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.656630] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquired lock "refresh_cache-fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.656786] env[61857]: DEBUG nova.network.neutron [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1130.730638] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951663, 'name': ReconfigVM_Task, 'duration_secs': 0.279573} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.730770] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 857977a6-b81b-4c94-b823-436e57b34ba0/2e0c8bb4-0827-486c-98ce-7994555ffc23-rescue.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1130.731577] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9ef66b-166c-4671-8d1a-34cef12ca0f6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.755958] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d7cbdf5-022d-48c2-8a69-bdf0ef7690ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.770971] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1130.770971] env[61857]: value = "task-951664" [ 1130.770971] env[61857]: _type = "Task" [ 1130.770971] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.780010] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951664, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.063825] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.064043] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.064232] env[61857]: DEBUG nova.network.neutron [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1131.187266] env[61857]: DEBUG nova.network.neutron [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1131.284422] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951664, 'name': ReconfigVM_Task, 'duration_secs': 0.140009} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.284753] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1131.285061] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e31e61ec-0916-4834-98cb-ed8a105e6720 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.291009] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1131.291009] env[61857]: value = "task-951665" [ 1131.291009] env[61857]: _type = "Task" [ 1131.291009] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.298511] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951665, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.310805] env[61857]: DEBUG nova.network.neutron [req-dd5112cf-86b5-426b-a5ac-c966fb01b164 req-9c48766f-9f51-46a6-b30a-0a31054087a0 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Updated VIF entry in instance network info cache for port 1c8b2449-1a4b-4062-a760-d88145f9615e. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1131.310916] env[61857]: DEBUG nova.network.neutron [req-dd5112cf-86b5-426b-a5ac-c966fb01b164 req-9c48766f-9f51-46a6-b30a-0a31054087a0 service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Updating instance_info_cache with network_info: [{"id": "1c8b2449-1a4b-4062-a760-d88145f9615e", "address": "fa:16:3e:b0:fd:68", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c8b2449-1a", "ovs_interfaceid": "1c8b2449-1a4b-4062-a760-d88145f9615e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.313142] env[61857]: DEBUG nova.network.neutron [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Updating instance_info_cache with network_info: [{"id": "9d257050-d1bb-4736-9db7-98fe7964a3e4", "address": "fa:16:3e:65:60:94", "network": {"id": "b9fee6b6-95da-4039-97d8-0f82c16a3a2a", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-341683080-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38a031bc6d084589a12d2246c95c0324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d257050-d1", "ovs_interfaceid": "9d257050-d1bb-4736-9db7-98fe7964a3e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.770015] env[61857]: DEBUG nova.network.neutron [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance_info_cache with network_info: [{"id": "f3163567-eee7-4192-9332-fe1748882d93", "address": "fa:16:3e:80:c4:35", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3163567-ee", "ovs_interfaceid": "f3163567-eee7-4192-9332-fe1748882d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.801234] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951665, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.813142] env[61857]: DEBUG oslo_concurrency.lockutils [req-dd5112cf-86b5-426b-a5ac-c966fb01b164 req-9c48766f-9f51-46a6-b30a-0a31054087a0 service nova] Releasing lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.814721] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Releasing lock "refresh_cache-fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.815011] env[61857]: DEBUG nova.compute.manager [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Instance network_info: |[{"id": "9d257050-d1bb-4736-9db7-98fe7964a3e4", "address": "fa:16:3e:65:60:94", "network": {"id": "b9fee6b6-95da-4039-97d8-0f82c16a3a2a", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-341683080-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38a031bc6d084589a12d2246c95c0324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d257050-d1", "ovs_interfaceid": "9d257050-d1bb-4736-9db7-98fe7964a3e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1131.815396] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:60:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c80243e-93a7-4a95-bc8d-e9534bacd66e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d257050-d1bb-4736-9db7-98fe7964a3e4', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1131.822887] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Creating folder: Project (38a031bc6d084589a12d2246c95c0324). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1131.823155] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-66c67d34-b599-4c4a-ae2a-81b78eddad88 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.833688] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Created folder: Project (38a031bc6d084589a12d2246c95c0324) in parent group-v214027. [ 1131.833892] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Creating folder: Instances. Parent ref: group-v214228. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1131.834166] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0284c84a-5048-4de3-a276-cd91c545b7bd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.847235] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Created folder: Instances in parent group-v214228. [ 1131.847478] env[61857]: DEBUG oslo.service.loopingcall [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1131.847673] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1131.847881] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3ec66c1-bb7a-4998-af3b-506efb982ede {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.866386] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1131.866386] env[61857]: value = "task-951668" [ 1131.866386] env[61857]: _type = "Task" [ 1131.866386] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.873612] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951668, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.076346] env[61857]: DEBUG nova.compute.manager [req-d1535e85-c143-4035-b260-e5db794bf076 req-2e722d29-c6c6-44ad-ae0d-05aead766a89 service nova] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Received event network-changed-9d257050-d1bb-4736-9db7-98fe7964a3e4 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1132.076559] env[61857]: DEBUG nova.compute.manager [req-d1535e85-c143-4035-b260-e5db794bf076 req-2e722d29-c6c6-44ad-ae0d-05aead766a89 service nova] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Refreshing instance network info cache due to event network-changed-9d257050-d1bb-4736-9db7-98fe7964a3e4. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1132.076899] env[61857]: DEBUG oslo_concurrency.lockutils [req-d1535e85-c143-4035-b260-e5db794bf076 req-2e722d29-c6c6-44ad-ae0d-05aead766a89 service nova] Acquiring lock "refresh_cache-fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.077174] env[61857]: DEBUG oslo_concurrency.lockutils [req-d1535e85-c143-4035-b260-e5db794bf076 req-2e722d29-c6c6-44ad-ae0d-05aead766a89 service nova] Acquired lock "refresh_cache-fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.077371] env[61857]: DEBUG nova.network.neutron [req-d1535e85-c143-4035-b260-e5db794bf076 req-2e722d29-c6c6-44ad-ae0d-05aead766a89 service nova] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Refreshing network info cache for port 9d257050-d1bb-4736-9db7-98fe7964a3e4 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1132.273055] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.301918] env[61857]: DEBUG oslo_vmware.api [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951665, 'name': PowerOnVM_Task, 'duration_secs': 0.57294} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.301918] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1132.304545] env[61857]: DEBUG nova.compute.manager [None req-b4e51933-bcc5-48dd-a6f0-e18d3a20b859 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1132.305341] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fed5cb7-54d9-435e-a9dd-e5c307158d06 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.375972] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951668, 'name': CreateVM_Task, 'duration_secs': 0.29329} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.376158] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1132.376839] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.377026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.377389] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1132.377632] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a195581-d7ee-4ec2-934b-8202f54a5095 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.382521] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for the task: (returnval){ [ 1132.382521] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5226051d-6746-15f0-1820-d0d00b1f2427" [ 1132.382521] env[61857]: _type = "Task" [ 1132.382521] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.393327] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5226051d-6746-15f0-1820-d0d00b1f2427, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.799878] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32abae7-2420-4d32-97f4-2f758ef27720 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.823813] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd430eb-ea79-419a-a0c2-544509cc9662 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.832625] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance '857dbf02-d829-4476-9094-13a30e14c799' progress to 83 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1132.861092] env[61857]: DEBUG nova.network.neutron [req-d1535e85-c143-4035-b260-e5db794bf076 req-2e722d29-c6c6-44ad-ae0d-05aead766a89 service nova] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Updated VIF entry in instance network info cache for port 9d257050-d1bb-4736-9db7-98fe7964a3e4. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1132.861487] env[61857]: DEBUG nova.network.neutron [req-d1535e85-c143-4035-b260-e5db794bf076 req-2e722d29-c6c6-44ad-ae0d-05aead766a89 service nova] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Updating instance_info_cache with network_info: [{"id": "9d257050-d1bb-4736-9db7-98fe7964a3e4", "address": "fa:16:3e:65:60:94", "network": {"id": "b9fee6b6-95da-4039-97d8-0f82c16a3a2a", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-341683080-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38a031bc6d084589a12d2246c95c0324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d257050-d1", "ovs_interfaceid": "9d257050-d1bb-4736-9db7-98fe7964a3e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.892703] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5226051d-6746-15f0-1820-d0d00b1f2427, 'name': SearchDatastore_Task, 'duration_secs': 0.009511} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.893025] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.893273] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1132.893510] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.893665] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.893850] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1132.894123] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23eff949-dfc8-4ac7-acaa-a9bf420f2140 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.902203] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1132.902427] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1132.903154] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4081c117-6d4e-4a2d-bcc9-877bed0bcdbc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.908036] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for the task: (returnval){ [ 1132.908036] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523d3206-b460-fc38-18a1-1a3b87d1bfab" [ 1132.908036] env[61857]: _type = "Task" [ 1132.908036] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.915143] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523d3206-b460-fc38-18a1-1a3b87d1bfab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.158123] env[61857]: DEBUG nova.compute.manager [req-d247bf59-7926-48ee-86dc-c1999911b022 req-eea81685-902f-485b-8d62-7d76c74df9e3 service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Received event network-changed-bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1133.158358] env[61857]: DEBUG nova.compute.manager [req-d247bf59-7926-48ee-86dc-c1999911b022 req-eea81685-902f-485b-8d62-7d76c74df9e3 service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Refreshing instance network info cache due to event network-changed-bfc4be35-7059-4d0f-bb2f-99c368eb3840. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1133.158570] env[61857]: DEBUG oslo_concurrency.lockutils [req-d247bf59-7926-48ee-86dc-c1999911b022 req-eea81685-902f-485b-8d62-7d76c74df9e3 service nova] Acquiring lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.158720] env[61857]: DEBUG oslo_concurrency.lockutils [req-d247bf59-7926-48ee-86dc-c1999911b022 req-eea81685-902f-485b-8d62-7d76c74df9e3 service nova] Acquired lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.158886] env[61857]: DEBUG nova.network.neutron [req-d247bf59-7926-48ee-86dc-c1999911b022 req-eea81685-902f-485b-8d62-7d76c74df9e3 service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Refreshing network info cache for port bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1133.341641] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9626bede-e7f7-4348-b263-2ec6175a1e61 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance '857dbf02-d829-4476-9094-13a30e14c799' progress to 100 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1133.363829] env[61857]: DEBUG oslo_concurrency.lockutils [req-d1535e85-c143-4035-b260-e5db794bf076 req-2e722d29-c6c6-44ad-ae0d-05aead766a89 service nova] Releasing lock "refresh_cache-fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.418854] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]523d3206-b460-fc38-18a1-1a3b87d1bfab, 'name': SearchDatastore_Task, 'duration_secs': 0.009704} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.420131] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33d2854b-4481-4ae7-89ac-004d50dbcb5e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.426025] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for the task: (returnval){ [ 1133.426025] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521baaec-5db6-0c68-f4ee-90b0a7adc0dd" [ 1133.426025] env[61857]: _type = "Task" [ 1133.426025] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.434078] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521baaec-5db6-0c68-f4ee-90b0a7adc0dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.938569] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]521baaec-5db6-0c68-f4ee-90b0a7adc0dd, 'name': SearchDatastore_Task, 'duration_secs': 0.009049} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.938959] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.939057] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] fa2b18b4-b5ba-4036-b50e-ccdefd1f0524/fa2b18b4-b5ba-4036-b50e-ccdefd1f0524.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1133.939323] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aa375677-be06-48d9-8a36-18eac289f746 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.946512] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for the task: (returnval){ [ 1133.946512] env[61857]: value = "task-951669" [ 1133.946512] env[61857]: _type = "Task" [ 1133.946512] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.954127] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951669, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.155613] env[61857]: DEBUG nova.network.neutron [req-d247bf59-7926-48ee-86dc-c1999911b022 req-eea81685-902f-485b-8d62-7d76c74df9e3 service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updated VIF entry in instance network info cache for port bfc4be35-7059-4d0f-bb2f-99c368eb3840. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1134.156124] env[61857]: DEBUG nova.network.neutron [req-d247bf59-7926-48ee-86dc-c1999911b022 req-eea81685-902f-485b-8d62-7d76c74df9e3 service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updating instance_info_cache with network_info: [{"id": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "address": "fa:16:3e:5d:01:97", "network": {"id": "dbaf4423-cac8-4e35-81f5-3a339a7a8299", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-321217147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f7ce5df26be54a4e8f258280a471f38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc4be35-70", "ovs_interfaceid": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.456551] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951669, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.392143} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.456826] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore1] fa2b18b4-b5ba-4036-b50e-ccdefd1f0524/fa2b18b4-b5ba-4036-b50e-ccdefd1f0524.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1134.457055] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1134.457320] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8506e7d4-19fe-4e96-a87d-9cbc8506a1d7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.463086] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for the task: (returnval){ [ 1134.463086] env[61857]: value = "task-951670" [ 1134.463086] env[61857]: _type = "Task" [ 1134.463086] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.470953] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951670, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.659162] env[61857]: DEBUG oslo_concurrency.lockutils [req-d247bf59-7926-48ee-86dc-c1999911b022 req-eea81685-902f-485b-8d62-7d76c74df9e3 service nova] Releasing lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.973488] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951670, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054485} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.973793] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1134.974566] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4dc3e8-1c3d-46b2-a5ba-4793fe9f6b0b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.996505] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] fa2b18b4-b5ba-4036-b50e-ccdefd1f0524/fa2b18b4-b5ba-4036-b50e-ccdefd1f0524.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.000018] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dde87437-39b2-481c-8ddc-97b592a2f8ca {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.018817] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for the task: (returnval){ [ 1135.018817] env[61857]: value = "task-951671" [ 1135.018817] env[61857]: _type = "Task" [ 1135.018817] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.026933] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951671, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.093855] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "857dbf02-d829-4476-9094-13a30e14c799" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.094163] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.094383] env[61857]: DEBUG nova.compute.manager [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Going to confirm migration 6 {{(pid=61857) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1135.191363] env[61857]: DEBUG nova.compute.manager [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Received event network-changed-bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1135.191557] env[61857]: DEBUG nova.compute.manager [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Refreshing instance network info cache due to event network-changed-bfc4be35-7059-4d0f-bb2f-99c368eb3840. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1135.191757] env[61857]: DEBUG oslo_concurrency.lockutils [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] Acquiring lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.191922] env[61857]: DEBUG oslo_concurrency.lockutils [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] Acquired lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.192096] env[61857]: DEBUG nova.network.neutron [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Refreshing network info cache for port bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1135.441524] env[61857]: DEBUG nova.compute.manager [req-f4303286-bdc9-4aee-8a5d-697842b67172 req-4fb59ac4-4521-46d5-a4ac-d9735fb5f054 service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Received event network-changed-bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1135.441751] env[61857]: DEBUG nova.compute.manager [req-f4303286-bdc9-4aee-8a5d-697842b67172 req-4fb59ac4-4521-46d5-a4ac-d9735fb5f054 service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Refreshing instance network info cache due to event network-changed-bfc4be35-7059-4d0f-bb2f-99c368eb3840. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1135.441947] env[61857]: DEBUG oslo_concurrency.lockutils [req-f4303286-bdc9-4aee-8a5d-697842b67172 req-4fb59ac4-4521-46d5-a4ac-d9735fb5f054 service nova] Acquiring lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.529592] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951671, 'name': ReconfigVM_Task, 'duration_secs': 0.266175} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.529878] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Reconfigured VM instance instance-0000006d to attach disk [datastore1] fa2b18b4-b5ba-4036-b50e-ccdefd1f0524/fa2b18b4-b5ba-4036-b50e-ccdefd1f0524.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1135.530562] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f1d0824-1eff-406c-9015-e9342d5a3760 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.537610] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for the task: (returnval){ [ 1135.537610] env[61857]: value = "task-951672" [ 1135.537610] env[61857]: _type = "Task" [ 1135.537610] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.550179] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951672, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.654177] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.654177] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.654338] env[61857]: DEBUG nova.network.neutron [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1135.654468] env[61857]: DEBUG nova.objects.instance [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'info_cache' on Instance uuid 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.927910] env[61857]: DEBUG nova.network.neutron [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updated VIF entry in instance network info cache for port bfc4be35-7059-4d0f-bb2f-99c368eb3840. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1135.928484] env[61857]: DEBUG nova.network.neutron [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updating instance_info_cache with network_info: [{"id": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "address": "fa:16:3e:5d:01:97", "network": {"id": "dbaf4423-cac8-4e35-81f5-3a339a7a8299", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-321217147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f7ce5df26be54a4e8f258280a471f38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc4be35-70", "ovs_interfaceid": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.048062] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951672, 'name': Rename_Task} progress is 14%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.431625] env[61857]: DEBUG oslo_concurrency.lockutils [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] Releasing lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.432101] env[61857]: DEBUG nova.compute.manager [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Received event network-changed-bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1136.432408] env[61857]: DEBUG nova.compute.manager [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Refreshing instance network info cache due to event network-changed-bfc4be35-7059-4d0f-bb2f-99c368eb3840. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1136.432743] env[61857]: DEBUG oslo_concurrency.lockutils [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] Acquiring lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.432988] env[61857]: DEBUG oslo_concurrency.lockutils [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] Acquired lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.433233] env[61857]: DEBUG nova.network.neutron [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Refreshing network info cache for port bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1136.547533] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951672, 'name': Rename_Task, 'duration_secs': 0.817246} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.547840] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1136.548104] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14fd1e17-8d50-4f19-ba7d-01f91985004c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.554931] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for the task: (returnval){ [ 1136.554931] env[61857]: value = "task-951673" [ 1136.554931] env[61857]: _type = "Task" [ 1136.554931] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.562391] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951673, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.586562] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquiring lock "857977a6-b81b-4c94-b823-436e57b34ba0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.586790] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lock "857977a6-b81b-4c94-b823-436e57b34ba0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.587028] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquiring lock "857977a6-b81b-4c94-b823-436e57b34ba0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.587229] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lock "857977a6-b81b-4c94-b823-436e57b34ba0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.587412] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lock "857977a6-b81b-4c94-b823-436e57b34ba0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.589652] env[61857]: INFO nova.compute.manager [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Terminating instance [ 1136.591537] env[61857]: DEBUG nova.compute.manager [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1136.591729] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1136.592586] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05368cd2-9093-46f9-9e9a-344fe9f6796d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.600811] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1136.601083] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ea70737-38d2-4851-84c2-14c6b940b416 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.608371] env[61857]: DEBUG oslo_vmware.api [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1136.608371] env[61857]: value = "task-951674" [ 1136.608371] env[61857]: _type = "Task" [ 1136.608371] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.615791] env[61857]: DEBUG oslo_vmware.api [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951674, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.913387] env[61857]: DEBUG nova.network.neutron [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance_info_cache with network_info: [{"id": "f3163567-eee7-4192-9332-fe1748882d93", "address": "fa:16:3e:80:c4:35", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3163567-ee", "ovs_interfaceid": "f3163567-eee7-4192-9332-fe1748882d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.064408] env[61857]: DEBUG oslo_vmware.api [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951673, 'name': PowerOnVM_Task, 'duration_secs': 0.435244} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.064692] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1137.064873] env[61857]: INFO nova.compute.manager [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Took 7.34 seconds to spawn the instance on the hypervisor. [ 1137.065070] env[61857]: DEBUG nova.compute.manager [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1137.065825] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8ff538-bfc7-44a6-8941-f4f43df8d1e4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.118247] env[61857]: DEBUG oslo_vmware.api [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951674, 'name': PowerOffVM_Task, 'duration_secs': 0.167174} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.118508] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1137.118686] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1137.118943] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e231f13-7356-4b1b-a36a-c8aed086007b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.131894] env[61857]: DEBUG nova.network.neutron [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updated VIF entry in instance network info cache for port bfc4be35-7059-4d0f-bb2f-99c368eb3840. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1137.132299] env[61857]: DEBUG nova.network.neutron [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updating instance_info_cache with network_info: [{"id": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "address": "fa:16:3e:5d:01:97", "network": {"id": "dbaf4423-cac8-4e35-81f5-3a339a7a8299", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-321217147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f7ce5df26be54a4e8f258280a471f38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc4be35-70", "ovs_interfaceid": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.183141] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1137.183396] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1137.183587] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Deleting the datastore file [datastore2] 857977a6-b81b-4c94-b823-436e57b34ba0 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1137.183855] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7e03d32-0a61-4fed-8bde-8b659bef2113 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.190140] env[61857]: DEBUG oslo_vmware.api [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for the task: (returnval){ [ 1137.190140] env[61857]: value = "task-951676" [ 1137.190140] env[61857]: _type = "Task" [ 1137.190140] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.198061] env[61857]: DEBUG oslo_vmware.api [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951676, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.416609] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.416944] env[61857]: DEBUG nova.objects.instance [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'migration_context' on Instance uuid 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.582067] env[61857]: INFO nova.compute.manager [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Took 13.08 seconds to build instance. [ 1137.635060] env[61857]: DEBUG oslo_concurrency.lockutils [req-76e74571-b2e5-4da8-ac52-a7a423b0ea97 req-614bb77a-3b43-4edd-8551-7966c4d5fe3a service nova] Releasing lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.635479] env[61857]: DEBUG oslo_concurrency.lockutils [req-f4303286-bdc9-4aee-8a5d-697842b67172 req-4fb59ac4-4521-46d5-a4ac-d9735fb5f054 service nova] Acquired lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.635696] env[61857]: DEBUG nova.network.neutron [req-f4303286-bdc9-4aee-8a5d-697842b67172 req-4fb59ac4-4521-46d5-a4ac-d9735fb5f054 service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Refreshing network info cache for port bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1137.700275] env[61857]: DEBUG oslo_vmware.api [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951676, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.799615] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquiring lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.919778] env[61857]: DEBUG nova.objects.base [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Object Instance<857dbf02-d829-4476-9094-13a30e14c799> lazy-loaded attributes: info_cache,migration_context {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1137.920735] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea41eb1b-29d0-4b04-b6e4-1365b78124d8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.941644] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82dea8c2-678d-4930-acc1-ccedca053f91 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.946999] env[61857]: DEBUG oslo_vmware.api [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1137.946999] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5256c706-6ce1-a9fb-3bf9-29a6ba8b38a2" [ 1137.946999] env[61857]: _type = "Task" [ 1137.946999] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.954702] env[61857]: DEBUG oslo_vmware.api [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5256c706-6ce1-a9fb-3bf9-29a6ba8b38a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.084585] env[61857]: DEBUG oslo_concurrency.lockutils [None req-10b9221e-23c0-47ae-a885-a37da3d7f26a tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.592s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.085158] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.285s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.085158] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquiring lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.085382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.085554] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.087845] env[61857]: INFO nova.compute.manager [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Terminating instance [ 1138.089860] env[61857]: DEBUG nova.compute.manager [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1138.090449] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1138.090976] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb64d806-d5c6-4ce9-a500-df275889e82c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.099209] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1138.099209] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60572b62-4e8a-40d8-b9ce-e92ebf665efe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.105365] env[61857]: DEBUG oslo_vmware.api [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for the task: (returnval){ [ 1138.105365] env[61857]: value = "task-951677" [ 1138.105365] env[61857]: _type = "Task" [ 1138.105365] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.113644] env[61857]: DEBUG oslo_vmware.api [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951677, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.200526] env[61857]: DEBUG oslo_vmware.api [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Task: {'id': task-951676, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.916365} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.200823] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.200964] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1138.201159] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1138.201342] env[61857]: INFO nova.compute.manager [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1138.201617] env[61857]: DEBUG oslo.service.loopingcall [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.201817] env[61857]: DEBUG nova.compute.manager [-] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1138.201912] env[61857]: DEBUG nova.network.neutron [-] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1138.324227] env[61857]: DEBUG nova.network.neutron [req-f4303286-bdc9-4aee-8a5d-697842b67172 req-4fb59ac4-4521-46d5-a4ac-d9735fb5f054 service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updated VIF entry in instance network info cache for port bfc4be35-7059-4d0f-bb2f-99c368eb3840. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1138.325061] env[61857]: DEBUG nova.network.neutron [req-f4303286-bdc9-4aee-8a5d-697842b67172 req-4fb59ac4-4521-46d5-a4ac-d9735fb5f054 service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updating instance_info_cache with network_info: [{"id": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "address": "fa:16:3e:5d:01:97", "network": {"id": "dbaf4423-cac8-4e35-81f5-3a339a7a8299", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-321217147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f7ce5df26be54a4e8f258280a471f38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfc4be35-70", "ovs_interfaceid": "bfc4be35-7059-4d0f-bb2f-99c368eb3840", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.458871] env[61857]: DEBUG oslo_vmware.api [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5256c706-6ce1-a9fb-3bf9-29a6ba8b38a2, 'name': SearchDatastore_Task, 'duration_secs': 0.015133} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.462014] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.462014] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.477365] env[61857]: DEBUG nova.compute.manager [req-65a8a48f-4b6a-4b5b-9a9a-b657b7b616d6 req-d80c3472-8630-4b9f-a1eb-3d74c9f9ee7c service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Received event network-vif-deleted-bfc4be35-7059-4d0f-bb2f-99c368eb3840 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1138.477602] env[61857]: INFO nova.compute.manager [req-65a8a48f-4b6a-4b5b-9a9a-b657b7b616d6 req-d80c3472-8630-4b9f-a1eb-3d74c9f9ee7c service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Neutron deleted interface bfc4be35-7059-4d0f-bb2f-99c368eb3840; detaching it from the instance and deleting it from the info cache [ 1138.477791] env[61857]: DEBUG nova.network.neutron [req-65a8a48f-4b6a-4b5b-9a9a-b657b7b616d6 req-d80c3472-8630-4b9f-a1eb-3d74c9f9ee7c service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.615761] env[61857]: DEBUG oslo_vmware.api [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951677, 'name': PowerOffVM_Task, 'duration_secs': 0.185027} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.616054] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1138.616234] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1138.616487] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63e9b15e-8ed2-4a20-86b5-6f7ee4c401af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.685030] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1138.685241] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1138.685446] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Deleting the datastore file [datastore1] fa2b18b4-b5ba-4036-b50e-ccdefd1f0524 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1138.685744] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f3ee325-1781-4bd1-a0e2-e275d84f1d19 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.691812] env[61857]: DEBUG oslo_vmware.api [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for the task: (returnval){ [ 1138.691812] env[61857]: value = "task-951679" [ 1138.691812] env[61857]: _type = "Task" [ 1138.691812] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.699517] env[61857]: DEBUG oslo_vmware.api [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951679, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.827263] env[61857]: DEBUG oslo_concurrency.lockutils [req-f4303286-bdc9-4aee-8a5d-697842b67172 req-4fb59ac4-4521-46d5-a4ac-d9735fb5f054 service nova] Releasing lock "refresh_cache-857977a6-b81b-4c94-b823-436e57b34ba0" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.940187] env[61857]: DEBUG nova.network.neutron [-] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.982479] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46bb8ce8-ce01-40bd-a8de-fe52fa6495c7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.991667] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e463923a-ca5c-4a9c-852f-201dbb47a1c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.019885] env[61857]: DEBUG nova.compute.manager [req-65a8a48f-4b6a-4b5b-9a9a-b657b7b616d6 req-d80c3472-8630-4b9f-a1eb-3d74c9f9ee7c service nova] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Detach interface failed, port_id=bfc4be35-7059-4d0f-bb2f-99c368eb3840, reason: Instance 857977a6-b81b-4c94-b823-436e57b34ba0 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1139.056489] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae59981-ccaf-46b0-9486-67a24884d0f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.063711] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d968753-df2e-4476-85a1-6dc365ef111c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.095760] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c176a6b-4952-4028-869f-43d49142442b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.103151] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3943f7f-afa0-4533-b8a1-e87a4485b9cd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.116085] env[61857]: DEBUG nova.compute.provider_tree [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.200926] env[61857]: DEBUG oslo_vmware.api [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Task: {'id': task-951679, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.118593} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.201157] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1139.201351] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1139.201640] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1139.201829] env[61857]: INFO nova.compute.manager [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1139.202088] env[61857]: DEBUG oslo.service.loopingcall [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1139.202286] env[61857]: DEBUG nova.compute.manager [-] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1139.202381] env[61857]: DEBUG nova.network.neutron [-] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1139.442462] env[61857]: INFO nova.compute.manager [-] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Took 1.24 seconds to deallocate network for instance. [ 1139.618826] env[61857]: DEBUG nova.scheduler.client.report [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1139.900947] env[61857]: DEBUG nova.network.neutron [-] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.948887] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.403402] env[61857]: INFO nova.compute.manager [-] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Took 1.20 seconds to deallocate network for instance. [ 1140.502131] env[61857]: DEBUG nova.compute.manager [req-87b52d08-5b5a-4573-b0cf-b78bffca5c77 req-880fab5f-15d0-47f2-bc73-7fe2635efd0a service nova] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Received event network-vif-deleted-9d257050-d1bb-4736-9db7-98fe7964a3e4 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1140.631456] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.172s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.631763] env[61857]: DEBUG nova.compute.manager [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61857) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1140.635163] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.686s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.635407] env[61857]: DEBUG nova.objects.instance [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lazy-loading 'resources' on Instance uuid 857977a6-b81b-4c94-b823-436e57b34ba0 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.909964] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.188289] env[61857]: INFO nova.scheduler.client.report [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleted allocation for migration 955e4413-8c36-40ff-8439-c037905e8eef [ 1141.213528] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a97590-a153-41d3-8a12-a542950f4c26 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.216621] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.222197] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ff4129-90c4-4eb8-8dab-3960b590badd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.251934] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172bb2cd-d04c-4ff8-b77a-31e07565b278 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.258837] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d99e5b2-2dd8-42aa-8626-1eb5ff05e408 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.271768] env[61857]: DEBUG nova.compute.provider_tree [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.684580] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.689196] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.694814] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bafb69dd-227e-4127-964e-3c047cdfb3cb tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.600s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.774319] env[61857]: DEBUG nova.scheduler.client.report [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1142.128572] env[61857]: DEBUG nova.objects.instance [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'flavor' on Instance uuid 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.279694] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.644s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.282141] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.372s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.282566] env[61857]: DEBUG nova.objects.instance [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lazy-loading 'resources' on Instance uuid fa2b18b4-b5ba-4036-b50e-ccdefd1f0524 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.297210] env[61857]: INFO nova.scheduler.client.report [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Deleted allocations for instance 857977a6-b81b-4c94-b823-436e57b34ba0 [ 1142.633679] env[61857]: DEBUG oslo_concurrency.lockutils [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.633910] env[61857]: DEBUG oslo_concurrency.lockutils [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.634197] env[61857]: DEBUG nova.network.neutron [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1142.634442] env[61857]: DEBUG nova.objects.instance [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'info_cache' on Instance uuid 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.806678] env[61857]: DEBUG oslo_concurrency.lockutils [None req-de9a4d3c-ad44-40c7-95c9-2e0f14828614 tempest-ServerRescueTestJSONUnderV235-412031647 tempest-ServerRescueTestJSONUnderV235-412031647-project-member] Lock "857977a6-b81b-4c94-b823-436e57b34ba0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.220s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.852019] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db97308-c620-4574-a5b8-1dda4b094d64 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.857397] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b157ac-c809-4b07-8038-07e10a4c8991 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.892204] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada4eda9-c2ab-40b9-92b2-0812d6daea77 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.901015] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae06257a-b229-40f4-b329-3da8549b8578 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.915047] env[61857]: DEBUG nova.compute.provider_tree [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.137811] env[61857]: DEBUG nova.objects.base [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Object Instance<857dbf02-d829-4476-9094-13a30e14c799> lazy-loaded attributes: flavor,info_cache {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1143.418483] env[61857]: DEBUG nova.scheduler.client.report [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1143.689601] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.689775] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1143.923373] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.952364] env[61857]: INFO nova.scheduler.client.report [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Deleted allocations for instance fa2b18b4-b5ba-4036-b50e-ccdefd1f0524 [ 1143.971986] env[61857]: DEBUG nova.network.neutron [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance_info_cache with network_info: [{"id": "f3163567-eee7-4192-9332-fe1748882d93", "address": "fa:16:3e:80:c4:35", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3163567-ee", "ovs_interfaceid": "f3163567-eee7-4192-9332-fe1748882d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.193630] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.465637] env[61857]: DEBUG oslo_concurrency.lockutils [None req-1d437fbd-463a-495c-8bd8-f7b1e7165eab tempest-ServerGroupTestJSON-918036920 tempest-ServerGroupTestJSON-918036920-project-member] Lock "fa2b18b4-b5ba-4036-b50e-ccdefd1f0524" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.381s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.475517] env[61857]: DEBUG oslo_concurrency.lockutils [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.475858] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquired lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.476065] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Forcefully refreshing network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1144.978604] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1144.981217] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8eb2b132-d4ce-4750-a345-c8f749233c35 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.989576] env[61857]: DEBUG oslo_vmware.api [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1144.989576] env[61857]: value = "task-951680" [ 1144.989576] env[61857]: _type = "Task" [ 1144.989576] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.998686] env[61857]: DEBUG oslo_vmware.api [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951680, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.498977] env[61857]: DEBUG oslo_vmware.api [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951680, 'name': PowerOnVM_Task, 'duration_secs': 0.388834} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.499242] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1145.499432] env[61857]: DEBUG nova.compute.manager [None req-95049808-95f8-4868-9fef-b1ce5c1afa6f tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1145.500202] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1a0376-152b-4ae4-93ca-6b01b5eec78e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.746636] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance_info_cache with network_info: [{"id": "f3163567-eee7-4192-9332-fe1748882d93", "address": "fa:16:3e:80:c4:35", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3163567-ee", "ovs_interfaceid": "f3163567-eee7-4192-9332-fe1748882d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.249432] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Releasing lock "refresh_cache-857dbf02-d829-4476-9094-13a30e14c799" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.249813] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updated the network info_cache for instance {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1146.249911] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.249986] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.250163] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.250970] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.754391] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.754760] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.754955] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.755137] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1146.756056] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa61963b-791b-491b-9d3c-2b49048fa096 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.764543] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570b8b61-af53-407d-bf68-fb7e02250545 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.778627] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29dc7f4b-9a9c-4cea-955d-3906f8a90451 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.784595] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc6ebdf4-f99c-42fc-99e1-86f6c05e7b58 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.812609] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180474MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1146.812758] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.812958] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.842468] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "857dbf02-d829-4476-9094-13a30e14c799" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.842719] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.842930] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "857dbf02-d829-4476-9094-13a30e14c799-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.843135] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.843311] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.845201] env[61857]: INFO nova.compute.manager [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Terminating instance [ 1146.846889] env[61857]: DEBUG nova.compute.manager [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1146.847103] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1146.847873] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0e5458-cb94-4d0f-90ea-2b2408aeed22 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.855376] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1146.855590] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-380ad81e-744c-440b-ae13-efeee7ce3b50 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.861657] env[61857]: DEBUG oslo_vmware.api [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1146.861657] env[61857]: value = "task-951681" [ 1146.861657] env[61857]: _type = "Task" [ 1146.861657] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.870414] env[61857]: DEBUG oslo_vmware.api [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951681, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.371412] env[61857]: DEBUG oslo_vmware.api [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951681, 'name': PowerOffVM_Task, 'duration_secs': 0.196177} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.371751] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1147.371794] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1147.372061] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd0861ff-7e8e-4244-a4cd-ac5f04d65fe8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.437958] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1147.438217] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1147.438431] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleting the datastore file [datastore2] 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1147.438716] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b1c04f7-e45f-4daf-bcef-7644a1ab2621 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.445149] env[61857]: DEBUG oslo_vmware.api [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1147.445149] env[61857]: value = "task-951683" [ 1147.445149] env[61857]: _type = "Task" [ 1147.445149] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.452924] env[61857]: DEBUG oslo_vmware.api [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951683, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.836742] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c6d20166-7251-406f-9003-b0599c2f5987 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1147.836910] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance b33cfca7-5621-47a8-bb84-59d82847d699 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1147.837050] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 857dbf02-d829-4476-9094-13a30e14c799 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1147.837231] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1147.837373] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1152MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1147.880413] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d0c5a3-0ff7-4856-93dd-113ee000e702 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.887894] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113fba1b-1c30-44c2-b3d6-6a45ad605429 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.916546] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9111ea56-c071-431c-888e-4bd1d203cb83 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.923335] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a18c181-1456-4ef4-b742-82cd6777986c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.935862] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.953300] env[61857]: DEBUG oslo_vmware.api [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951683, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153036} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.953537] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1147.953724] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1147.953904] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1147.954105] env[61857]: INFO nova.compute.manager [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1147.954349] env[61857]: DEBUG oslo.service.loopingcall [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1147.954541] env[61857]: DEBUG nova.compute.manager [-] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1147.954639] env[61857]: DEBUG nova.network.neutron [-] [instance: 857dbf02-d829-4476-9094-13a30e14c799] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1148.438772] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1148.542726] env[61857]: DEBUG nova.compute.manager [req-9bb0432a-f0c9-4bb1-a922-c91e11fd1840 req-0c064019-6802-4b91-a560-20d2435eb2db service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Received event network-vif-deleted-f3163567-eee7-4192-9332-fe1748882d93 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1148.542941] env[61857]: INFO nova.compute.manager [req-9bb0432a-f0c9-4bb1-a922-c91e11fd1840 req-0c064019-6802-4b91-a560-20d2435eb2db service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Neutron deleted interface f3163567-eee7-4192-9332-fe1748882d93; detaching it from the instance and deleting it from the info cache [ 1148.543134] env[61857]: DEBUG nova.network.neutron [req-9bb0432a-f0c9-4bb1-a922-c91e11fd1840 req-0c064019-6802-4b91-a560-20d2435eb2db service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.945804] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1148.945804] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.133s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.018466] env[61857]: DEBUG nova.network.neutron [-] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.046183] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf50a3bb-f880-4750-8aff-788a6ff1f613 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.055481] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2b11af-63a3-467a-b843-50bae4342494 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.081055] env[61857]: DEBUG nova.compute.manager [req-9bb0432a-f0c9-4bb1-a922-c91e11fd1840 req-0c064019-6802-4b91-a560-20d2435eb2db service nova] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Detach interface failed, port_id=f3163567-eee7-4192-9332-fe1748882d93, reason: Instance 857dbf02-d829-4476-9094-13a30e14c799 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1149.520836] env[61857]: INFO nova.compute.manager [-] [instance: 857dbf02-d829-4476-9094-13a30e14c799] Took 1.57 seconds to deallocate network for instance. [ 1150.027391] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.027715] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.027946] env[61857]: DEBUG nova.objects.instance [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'resources' on Instance uuid 857dbf02-d829-4476-9094-13a30e14c799 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1150.385086] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.385290] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1150.583674] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9157c04f-598d-471e-a5a6-8a4b93388a03 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.591243] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88337bea-4935-4bf2-a4a4-43fe322c23db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.621570] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f6f88c-8764-4190-8560-467a3c4c2db4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.628422] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a27cc0-ea00-49e7-95a5-025e6edc48d2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.641848] env[61857]: DEBUG nova.compute.provider_tree [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1151.144993] env[61857]: DEBUG nova.scheduler.client.report [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1151.650064] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.622s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.668077] env[61857]: INFO nova.scheduler.client.report [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleted allocations for instance 857dbf02-d829-4476-9094-13a30e14c799 [ 1152.175966] env[61857]: DEBUG oslo_concurrency.lockutils [None req-6529fe37-0da8-4daa-86aa-e4fe78aca91a tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "857dbf02-d829-4476-9094-13a30e14c799" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.333s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.820562] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "299da51d-22a9-4da5-992a-482b63b81034" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.820848] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.323765] env[61857]: DEBUG nova.compute.manager [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1153.847360] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.847722] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.849693] env[61857]: INFO nova.compute.claims [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1154.902336] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9384af4a-4091-4067-9b77-a38fc0c6ad5f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.910242] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e339cef8-8ecf-45df-a4ad-582d6af581f2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.939492] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afaa1ba2-f2c3-4625-856b-afbd40a06e5f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.946550] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05302c5c-2f1a-4875-b711-a533436bfe5a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.959160] env[61857]: DEBUG nova.compute.provider_tree [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.462410] env[61857]: DEBUG nova.scheduler.client.report [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1155.967293] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.967869] env[61857]: DEBUG nova.compute.manager [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1156.473087] env[61857]: DEBUG nova.compute.utils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1156.474596] env[61857]: DEBUG nova.compute.manager [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1156.474675] env[61857]: DEBUG nova.network.neutron [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1156.529669] env[61857]: DEBUG nova.policy [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4c9ee17921646978e8b21ad5508813c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2a7d2f228f88485f863b3c421b25a37c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1156.810194] env[61857]: DEBUG nova.network.neutron [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Successfully created port: c6a429a1-74cb-4ee8-8024-7d25225c2265 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1156.977812] env[61857]: DEBUG nova.compute.manager [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1157.987334] env[61857]: DEBUG nova.compute.manager [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1158.012085] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1158.012358] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1158.012783] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1158.012783] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1158.012894] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1158.013095] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1158.013362] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1158.013498] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1158.013673] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1158.013843] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1158.014032] env[61857]: DEBUG nova.virt.hardware [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1158.014926] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606d29fa-7e31-4be8-81dc-320cf591cf45 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.023218] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b07f86-0a79-41aa-836e-fa8d34d6228b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.175591] env[61857]: DEBUG nova.compute.manager [req-ec455cd6-1621-457c-b326-0cbd3741d5ed req-de81c644-296c-4596-829b-658558aef57e service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Received event network-vif-plugged-c6a429a1-74cb-4ee8-8024-7d25225c2265 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1158.175826] env[61857]: DEBUG oslo_concurrency.lockutils [req-ec455cd6-1621-457c-b326-0cbd3741d5ed req-de81c644-296c-4596-829b-658558aef57e service nova] Acquiring lock "299da51d-22a9-4da5-992a-482b63b81034-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.176058] env[61857]: DEBUG oslo_concurrency.lockutils [req-ec455cd6-1621-457c-b326-0cbd3741d5ed req-de81c644-296c-4596-829b-658558aef57e service nova] Lock "299da51d-22a9-4da5-992a-482b63b81034-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.176239] env[61857]: DEBUG oslo_concurrency.lockutils [req-ec455cd6-1621-457c-b326-0cbd3741d5ed req-de81c644-296c-4596-829b-658558aef57e service nova] Lock "299da51d-22a9-4da5-992a-482b63b81034-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.176418] env[61857]: DEBUG nova.compute.manager [req-ec455cd6-1621-457c-b326-0cbd3741d5ed req-de81c644-296c-4596-829b-658558aef57e service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] No waiting events found dispatching network-vif-plugged-c6a429a1-74cb-4ee8-8024-7d25225c2265 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1158.176593] env[61857]: WARNING nova.compute.manager [req-ec455cd6-1621-457c-b326-0cbd3741d5ed req-de81c644-296c-4596-829b-658558aef57e service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Received unexpected event network-vif-plugged-c6a429a1-74cb-4ee8-8024-7d25225c2265 for instance with vm_state building and task_state spawning. [ 1158.254641] env[61857]: DEBUG nova.network.neutron [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Successfully updated port: c6a429a1-74cb-4ee8-8024-7d25225c2265 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1158.758060] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.758060] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.758060] env[61857]: DEBUG nova.network.neutron [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1159.288907] env[61857]: DEBUG nova.network.neutron [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1159.408468] env[61857]: DEBUG nova.network.neutron [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance_info_cache with network_info: [{"id": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "address": "fa:16:3e:0b:95:f2", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6a429a1-74", "ovs_interfaceid": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.911459] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.911821] env[61857]: DEBUG nova.compute.manager [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Instance network_info: |[{"id": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "address": "fa:16:3e:0b:95:f2", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6a429a1-74", "ovs_interfaceid": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1159.912283] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:95:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '170f3b82-5915-4e36-bce9-4664ebb6be5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6a429a1-74cb-4ee8-8024-7d25225c2265', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1159.919778] env[61857]: DEBUG oslo.service.loopingcall [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1159.919998] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1159.920237] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-727b8ea3-698a-416e-9aff-75252d516baa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.940078] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1159.940078] env[61857]: value = "task-951684" [ 1159.940078] env[61857]: _type = "Task" [ 1159.940078] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.947342] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951684, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.273250] env[61857]: DEBUG nova.compute.manager [req-3b3ac662-861f-4a8a-9844-3ea92e2031a5 req-0c5aadc9-beb0-4cef-975e-391fbb579d13 service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Received event network-changed-c6a429a1-74cb-4ee8-8024-7d25225c2265 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1160.273250] env[61857]: DEBUG nova.compute.manager [req-3b3ac662-861f-4a8a-9844-3ea92e2031a5 req-0c5aadc9-beb0-4cef-975e-391fbb579d13 service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Refreshing instance network info cache due to event network-changed-c6a429a1-74cb-4ee8-8024-7d25225c2265. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1160.273250] env[61857]: DEBUG oslo_concurrency.lockutils [req-3b3ac662-861f-4a8a-9844-3ea92e2031a5 req-0c5aadc9-beb0-4cef-975e-391fbb579d13 service nova] Acquiring lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.273386] env[61857]: DEBUG oslo_concurrency.lockutils [req-3b3ac662-861f-4a8a-9844-3ea92e2031a5 req-0c5aadc9-beb0-4cef-975e-391fbb579d13 service nova] Acquired lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.273558] env[61857]: DEBUG nova.network.neutron [req-3b3ac662-861f-4a8a-9844-3ea92e2031a5 req-0c5aadc9-beb0-4cef-975e-391fbb579d13 service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Refreshing network info cache for port c6a429a1-74cb-4ee8-8024-7d25225c2265 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1160.449631] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951684, 'name': CreateVM_Task, 'duration_secs': 0.294372} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.451036] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1160.451036] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.451214] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.451471] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1160.451732] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af6eca3f-944e-4bd6-8843-08a95e23327f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.455992] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1160.455992] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]522847c4-94aa-b79d-2c72-e054593deb94" [ 1160.455992] env[61857]: _type = "Task" [ 1160.455992] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.464521] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522847c4-94aa-b79d-2c72-e054593deb94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.965763] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522847c4-94aa-b79d-2c72-e054593deb94, 'name': SearchDatastore_Task, 'duration_secs': 0.011435} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.966066] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.966304] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1160.966540] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.966694] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.966876] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1160.967148] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a5aef77-4a33-49dc-9112-9e21d78a6d72 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.974478] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1160.974661] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1160.975342] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-806a984b-5a12-4b27-89f1-6328367d9117 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.981964] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1160.981964] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52ab2d0c-e392-a571-00df-479cc268c925" [ 1160.981964] env[61857]: _type = "Task" [ 1160.981964] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.989139] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ab2d0c-e392-a571-00df-479cc268c925, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.013491] env[61857]: DEBUG nova.network.neutron [req-3b3ac662-861f-4a8a-9844-3ea92e2031a5 req-0c5aadc9-beb0-4cef-975e-391fbb579d13 service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updated VIF entry in instance network info cache for port c6a429a1-74cb-4ee8-8024-7d25225c2265. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1161.013844] env[61857]: DEBUG nova.network.neutron [req-3b3ac662-861f-4a8a-9844-3ea92e2031a5 req-0c5aadc9-beb0-4cef-975e-391fbb579d13 service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance_info_cache with network_info: [{"id": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "address": "fa:16:3e:0b:95:f2", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6a429a1-74", "ovs_interfaceid": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.492472] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52ab2d0c-e392-a571-00df-479cc268c925, 'name': SearchDatastore_Task, 'duration_secs': 0.008431} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.493252] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98c719f7-6475-45ca-adb3-fc6fe816aa61 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.498082] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1161.498082] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5280dfec-66d8-6cde-c015-0e0f508fea91" [ 1161.498082] env[61857]: _type = "Task" [ 1161.498082] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.505097] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5280dfec-66d8-6cde-c015-0e0f508fea91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.516697] env[61857]: DEBUG oslo_concurrency.lockutils [req-3b3ac662-861f-4a8a-9844-3ea92e2031a5 req-0c5aadc9-beb0-4cef-975e-391fbb579d13 service nova] Releasing lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.008854] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5280dfec-66d8-6cde-c015-0e0f508fea91, 'name': SearchDatastore_Task, 'duration_secs': 0.009164} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.009154] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.009418] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 299da51d-22a9-4da5-992a-482b63b81034/299da51d-22a9-4da5-992a-482b63b81034.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1162.009682] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-431e6fb8-8be9-4279-91e6-c6b716937ac5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.016780] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1162.016780] env[61857]: value = "task-951685" [ 1162.016780] env[61857]: _type = "Task" [ 1162.016780] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.024179] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.262812] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "c6d20166-7251-406f-9003-b0599c2f5987" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.263120] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "c6d20166-7251-406f-9003-b0599c2f5987" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.263469] env[61857]: INFO nova.compute.manager [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Shelving [ 1162.527251] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951685, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.406606} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.527584] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 299da51d-22a9-4da5-992a-482b63b81034/299da51d-22a9-4da5-992a-482b63b81034.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1162.527688] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1162.527939] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8b69a67-35fc-4cb0-8617-ba566837208b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.534157] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1162.534157] env[61857]: value = "task-951686" [ 1162.534157] env[61857]: _type = "Task" [ 1162.534157] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.541825] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951686, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.771190] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1162.771464] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-286c6d33-3ad2-46fb-ac44-88c6f272070d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.778232] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1162.778232] env[61857]: value = "task-951687" [ 1162.778232] env[61857]: _type = "Task" [ 1162.778232] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.785749] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951687, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.044872] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951686, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069151} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.045137] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1163.045867] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b3a929-d677-4d65-98cc-abc3f15510c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.066957] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 299da51d-22a9-4da5-992a-482b63b81034/299da51d-22a9-4da5-992a-482b63b81034.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1163.067201] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40f1f434-e4ae-4971-8228-9e45424101f3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.086101] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1163.086101] env[61857]: value = "task-951688" [ 1163.086101] env[61857]: _type = "Task" [ 1163.086101] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.093286] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951688, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.287789] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951687, 'name': PowerOffVM_Task, 'duration_secs': 0.17952} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.288090] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1163.288890] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8de82b-d3ab-48f3-a4fe-956ad701d748 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.306900] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d823a0b-3009-4d16-9269-b289be0aa35f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.595950] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951688, 'name': ReconfigVM_Task, 'duration_secs': 0.273973} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.596308] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 299da51d-22a9-4da5-992a-482b63b81034/299da51d-22a9-4da5-992a-482b63b81034.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.596860] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10ff1a5c-7db1-466d-84a0-8d2fe7bae17f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.603612] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1163.603612] env[61857]: value = "task-951689" [ 1163.603612] env[61857]: _type = "Task" [ 1163.603612] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.610853] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951689, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.817169] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1163.817512] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cf9ebdac-9122-4bb2-a08a-3537f46be4d1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.825167] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1163.825167] env[61857]: value = "task-951690" [ 1163.825167] env[61857]: _type = "Task" [ 1163.825167] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.833095] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951690, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.113275] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951689, 'name': Rename_Task, 'duration_secs': 0.135033} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.113604] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1164.113864] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e8e2e81-cfe2-4867-954b-69959c246979 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.120204] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1164.120204] env[61857]: value = "task-951691" [ 1164.120204] env[61857]: _type = "Task" [ 1164.120204] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.127966] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951691, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.335836] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951690, 'name': CreateSnapshot_Task, 'duration_secs': 0.397735} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.336125] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1164.336850] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb2bb86-29c4-4d93-b022-142ff849bcd8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.635068] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951691, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.854098] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1164.854433] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-289ac14b-7aa1-4ce9-84c0-59f73326641b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.863185] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1164.863185] env[61857]: value = "task-951692" [ 1164.863185] env[61857]: _type = "Task" [ 1164.863185] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.872403] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951692, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.131371] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951691, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.373744] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951692, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.631724] env[61857]: DEBUG oslo_vmware.api [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951691, 'name': PowerOnVM_Task, 'duration_secs': 1.082012} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.632013] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1165.632268] env[61857]: INFO nova.compute.manager [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Took 7.64 seconds to spawn the instance on the hypervisor. [ 1165.632461] env[61857]: DEBUG nova.compute.manager [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1165.633293] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8d1d8a-4ce7-48b8-95be-748b0790d1e0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.874970] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951692, 'name': CloneVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.151746] env[61857]: INFO nova.compute.manager [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Took 12.32 seconds to build instance. [ 1166.380075] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951692, 'name': CloneVM_Task, 'duration_secs': 1.011871} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.380565] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Created linked-clone VM from snapshot [ 1166.381776] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2784c2b-4088-4393-96b2-8702c431c9b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.392746] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Uploading image 16cdeabe-6880-4e6e-9892-9af0bc97d9b3 {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1166.414945] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1166.414945] env[61857]: value = "vm-214233" [ 1166.414945] env[61857]: _type = "VirtualMachine" [ 1166.414945] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1166.415249] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ddbd9bb2-0a1b-47b6-b743-6c3d77cca982 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.426753] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lease: (returnval){ [ 1166.426753] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523a0ee1-4ac2-8408-7682-39c403038777" [ 1166.426753] env[61857]: _type = "HttpNfcLease" [ 1166.426753] env[61857]: } obtained for exporting VM: (result){ [ 1166.426753] env[61857]: value = "vm-214233" [ 1166.426753] env[61857]: _type = "VirtualMachine" [ 1166.426753] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1166.427031] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the lease: (returnval){ [ 1166.427031] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523a0ee1-4ac2-8408-7682-39c403038777" [ 1166.427031] env[61857]: _type = "HttpNfcLease" [ 1166.427031] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1166.433316] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1166.433316] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523a0ee1-4ac2-8408-7682-39c403038777" [ 1166.433316] env[61857]: _type = "HttpNfcLease" [ 1166.433316] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1166.654287] env[61857]: DEBUG oslo_concurrency.lockutils [None req-7ff2050c-a5ee-46f2-9646-06dbab6baaaa tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.833s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.935261] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1166.935261] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523a0ee1-4ac2-8408-7682-39c403038777" [ 1166.935261] env[61857]: _type = "HttpNfcLease" [ 1166.935261] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1166.935689] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1166.935689] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]523a0ee1-4ac2-8408-7682-39c403038777" [ 1166.935689] env[61857]: _type = "HttpNfcLease" [ 1166.935689] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1166.936276] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78dc41ef-f07b-421f-abab-114b9eb2a530 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.943241] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52668228-5cc9-7629-eb4f-2fddb9719143/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1166.943426] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52668228-5cc9-7629-eb4f-2fddb9719143/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1167.035343] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0eb041b3-0d7d-43d3-b5c0-7edaf538ed91 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.053150] env[61857]: DEBUG nova.compute.manager [req-9ba74459-1904-40a7-a543-4614fb146d30 req-f205ab65-cc17-41b5-b2ce-6e295c0350c9 service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Received event network-changed-c6a429a1-74cb-4ee8-8024-7d25225c2265 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1167.053150] env[61857]: DEBUG nova.compute.manager [req-9ba74459-1904-40a7-a543-4614fb146d30 req-f205ab65-cc17-41b5-b2ce-6e295c0350c9 service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Refreshing instance network info cache due to event network-changed-c6a429a1-74cb-4ee8-8024-7d25225c2265. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1167.053150] env[61857]: DEBUG oslo_concurrency.lockutils [req-9ba74459-1904-40a7-a543-4614fb146d30 req-f205ab65-cc17-41b5-b2ce-6e295c0350c9 service nova] Acquiring lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.053150] env[61857]: DEBUG oslo_concurrency.lockutils [req-9ba74459-1904-40a7-a543-4614fb146d30 req-f205ab65-cc17-41b5-b2ce-6e295c0350c9 service nova] Acquired lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.053150] env[61857]: DEBUG nova.network.neutron [req-9ba74459-1904-40a7-a543-4614fb146d30 req-f205ab65-cc17-41b5-b2ce-6e295c0350c9 service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Refreshing network info cache for port c6a429a1-74cb-4ee8-8024-7d25225c2265 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1167.574939] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "b33cfca7-5621-47a8-bb84-59d82847d699" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.574939] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "b33cfca7-5621-47a8-bb84-59d82847d699" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.574939] env[61857]: DEBUG nova.compute.manager [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1167.574939] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0694f80-00f8-4247-8fba-8cfe6817dd6b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.582634] env[61857]: DEBUG nova.compute.manager [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61857) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1167.583278] env[61857]: DEBUG nova.objects.instance [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'flavor' on Instance uuid b33cfca7-5621-47a8-bb84-59d82847d699 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.782320] env[61857]: DEBUG nova.network.neutron [req-9ba74459-1904-40a7-a543-4614fb146d30 req-f205ab65-cc17-41b5-b2ce-6e295c0350c9 service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updated VIF entry in instance network info cache for port c6a429a1-74cb-4ee8-8024-7d25225c2265. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1167.782846] env[61857]: DEBUG nova.network.neutron [req-9ba74459-1904-40a7-a543-4614fb146d30 req-f205ab65-cc17-41b5-b2ce-6e295c0350c9 service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance_info_cache with network_info: [{"id": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "address": "fa:16:3e:0b:95:f2", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6a429a1-74", "ovs_interfaceid": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.089564] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1168.090062] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-846d68bc-6707-4bbf-95ec-5aa22018e064 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.098591] env[61857]: DEBUG oslo_vmware.api [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1168.098591] env[61857]: value = "task-951694" [ 1168.098591] env[61857]: _type = "Task" [ 1168.098591] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.108371] env[61857]: DEBUG oslo_vmware.api [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951694, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.286720] env[61857]: DEBUG oslo_concurrency.lockutils [req-9ba74459-1904-40a7-a543-4614fb146d30 req-f205ab65-cc17-41b5-b2ce-6e295c0350c9 service nova] Releasing lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.610238] env[61857]: DEBUG oslo_vmware.api [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951694, 'name': PowerOffVM_Task, 'duration_secs': 0.267957} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.610664] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1168.610912] env[61857]: DEBUG nova.compute.manager [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1168.611852] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5366204-984b-4801-a241-2e69693e6fb2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.123514] env[61857]: DEBUG oslo_concurrency.lockutils [None req-0c8aaa19-f963-41a5-9322-c1d4ff6c1d7f tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "b33cfca7-5621-47a8-bb84-59d82847d699" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.550s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.988600] env[61857]: DEBUG nova.objects.instance [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'flavor' on Instance uuid b33cfca7-5621-47a8-bb84-59d82847d699 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1170.496088] env[61857]: DEBUG oslo_concurrency.lockutils [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.496088] env[61857]: DEBUG oslo_concurrency.lockutils [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.496088] env[61857]: DEBUG nova.network.neutron [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1170.496088] env[61857]: DEBUG nova.objects.instance [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'info_cache' on Instance uuid b33cfca7-5621-47a8-bb84-59d82847d699 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.000717] env[61857]: DEBUG nova.objects.base [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1171.727132] env[61857]: DEBUG nova.network.neutron [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Updating instance_info_cache with network_info: [{"id": "1c8b2449-1a4b-4062-a760-d88145f9615e", "address": "fa:16:3e:b0:fd:68", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c8b2449-1a", "ovs_interfaceid": "1c8b2449-1a4b-4062-a760-d88145f9615e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.230020] env[61857]: DEBUG oslo_concurrency.lockutils [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.733869] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1172.734279] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-541dd5f0-d910-480b-b752-6258c6c50275 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.742636] env[61857]: DEBUG oslo_vmware.api [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1172.742636] env[61857]: value = "task-951695" [ 1172.742636] env[61857]: _type = "Task" [ 1172.742636] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.750880] env[61857]: DEBUG oslo_vmware.api [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951695, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.252668] env[61857]: DEBUG oslo_vmware.api [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951695, 'name': PowerOnVM_Task, 'duration_secs': 0.445854} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.252668] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1173.252806] env[61857]: DEBUG nova.compute.manager [None req-64f1bd0c-019a-42b4-9f26-dc13ce0fd123 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1173.253608] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a541d404-a88b-4c4b-aca0-0ad9e7731149 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.043360] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0441b1b0-6994-4a4f-ad54-af5466652621 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.050300] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-91e745bc-3563-4e0d-9fd7-a80834826211 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Suspending the VM {{(pid=61857) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1175.050579] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-b544ed36-9721-4329-a21b-c12f6d69b519 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.057308] env[61857]: DEBUG oslo_vmware.api [None req-91e745bc-3563-4e0d-9fd7-a80834826211 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1175.057308] env[61857]: value = "task-951696" [ 1175.057308] env[61857]: _type = "Task" [ 1175.057308] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.065418] env[61857]: DEBUG oslo_vmware.api [None req-91e745bc-3563-4e0d-9fd7-a80834826211 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951696, 'name': SuspendVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.568094] env[61857]: DEBUG oslo_vmware.api [None req-91e745bc-3563-4e0d-9fd7-a80834826211 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951696, 'name': SuspendVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.068403] env[61857]: DEBUG oslo_vmware.api [None req-91e745bc-3563-4e0d-9fd7-a80834826211 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951696, 'name': SuspendVM_Task, 'duration_secs': 0.640584} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.068802] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-91e745bc-3563-4e0d-9fd7-a80834826211 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Suspended the VM {{(pid=61857) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1176.068930] env[61857]: DEBUG nova.compute.manager [None req-91e745bc-3563-4e0d-9fd7-a80834826211 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1176.069756] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013a5c80-e01e-4434-8165-c22828d140b7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.672291] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52668228-5cc9-7629-eb4f-2fddb9719143/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1176.673289] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4a414b-20b1-4c96-a197-e676a199f22e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.679411] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52668228-5cc9-7629-eb4f-2fddb9719143/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1176.679576] env[61857]: ERROR oslo_vmware.rw_handles [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52668228-5cc9-7629-eb4f-2fddb9719143/disk-0.vmdk due to incomplete transfer. [ 1176.679793] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-44e866ea-d4a3-498e-b865-852383812d52 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.685793] env[61857]: DEBUG oslo_vmware.rw_handles [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52668228-5cc9-7629-eb4f-2fddb9719143/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1176.685998] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Uploaded image 16cdeabe-6880-4e6e-9892-9af0bc97d9b3 to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1176.688193] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1176.688434] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-235908a6-e2eb-47d6-88c3-3ed05a04d2d5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.693705] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1176.693705] env[61857]: value = "task-951697" [ 1176.693705] env[61857]: _type = "Task" [ 1176.693705] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.701133] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951697, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.203266] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951697, 'name': Destroy_Task, 'duration_secs': 0.311926} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.203609] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Destroyed the VM [ 1177.203798] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1177.204113] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-17008c1e-0c6e-43a6-b8e7-e2a2aa49e363 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.209604] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1177.209604] env[61857]: value = "task-951698" [ 1177.209604] env[61857]: _type = "Task" [ 1177.209604] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.216773] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951698, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.401773] env[61857]: INFO nova.compute.manager [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Resuming [ 1177.402566] env[61857]: DEBUG nova.objects.instance [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'flavor' on Instance uuid b33cfca7-5621-47a8-bb84-59d82847d699 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.719063] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951698, 'name': RemoveSnapshot_Task, 'duration_secs': 0.423731} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.719361] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1177.719638] env[61857]: DEBUG nova.compute.manager [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1177.720425] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7745f0-2ddb-4a70-b394-017f33979266 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.232512] env[61857]: INFO nova.compute.manager [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Shelve offloading [ 1178.234513] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1178.234765] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-475441a0-1c83-4867-b1ba-8852c5b848ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.242279] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1178.242279] env[61857]: value = "task-951699" [ 1178.242279] env[61857]: _type = "Task" [ 1178.242279] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.251078] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] VM already powered off {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1178.251262] env[61857]: DEBUG nova.compute.manager [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1178.252017] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4b31ec-8570-40f2-8b2a-881be3898284 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.257159] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.257335] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.257511] env[61857]: DEBUG nova.network.neutron [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1178.410085] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.410301] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquired lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.410518] env[61857]: DEBUG nova.network.neutron [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1178.964229] env[61857]: DEBUG nova.network.neutron [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updating instance_info_cache with network_info: [{"id": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "address": "fa:16:3e:91:72:71", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a0b5327-2d", "ovs_interfaceid": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.103608] env[61857]: DEBUG nova.network.neutron [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Updating instance_info_cache with network_info: [{"id": "1c8b2449-1a4b-4062-a760-d88145f9615e", "address": "fa:16:3e:b0:fd:68", "network": {"id": "a5c49adf-421d-4cc6-baee-d372d9634955", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1936064613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ef111c8b3ff4f2383b4e81db026792b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c8b2449-1a", "ovs_interfaceid": "1c8b2449-1a4b-4062-a760-d88145f9615e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.471070] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.606908] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Releasing lock "refresh_cache-b33cfca7-5621-47a8-bb84-59d82847d699" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.607924] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ad0bb2-3ffb-4046-a573-c438293a3853 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.614390] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Resuming the VM {{(pid=61857) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1179.614626] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b354a41-894b-42e0-8a64-7cfb0b0353e7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.621610] env[61857]: DEBUG oslo_vmware.api [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1179.621610] env[61857]: value = "task-951700" [ 1179.621610] env[61857]: _type = "Task" [ 1179.621610] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.629027] env[61857]: DEBUG oslo_vmware.api [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951700, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.740138] env[61857]: DEBUG nova.compute.manager [req-d380e125-8067-4085-9919-3bc8a04363a2 req-cbea0003-ee69-4225-b8a0-f6924f5d1b6a service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received event network-vif-unplugged-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1179.740418] env[61857]: DEBUG oslo_concurrency.lockutils [req-d380e125-8067-4085-9919-3bc8a04363a2 req-cbea0003-ee69-4225-b8a0-f6924f5d1b6a service nova] Acquiring lock "c6d20166-7251-406f-9003-b0599c2f5987-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.740653] env[61857]: DEBUG oslo_concurrency.lockutils [req-d380e125-8067-4085-9919-3bc8a04363a2 req-cbea0003-ee69-4225-b8a0-f6924f5d1b6a service nova] Lock "c6d20166-7251-406f-9003-b0599c2f5987-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.740832] env[61857]: DEBUG oslo_concurrency.lockutils [req-d380e125-8067-4085-9919-3bc8a04363a2 req-cbea0003-ee69-4225-b8a0-f6924f5d1b6a service nova] Lock "c6d20166-7251-406f-9003-b0599c2f5987-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.741280] env[61857]: DEBUG nova.compute.manager [req-d380e125-8067-4085-9919-3bc8a04363a2 req-cbea0003-ee69-4225-b8a0-f6924f5d1b6a service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] No waiting events found dispatching network-vif-unplugged-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1179.741609] env[61857]: WARNING nova.compute.manager [req-d380e125-8067-4085-9919-3bc8a04363a2 req-cbea0003-ee69-4225-b8a0-f6924f5d1b6a service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received unexpected event network-vif-unplugged-9a0b5327-2d9d-47ef-bb93-6900b47fae64 for instance with vm_state shelved and task_state shelving_offloading. [ 1179.851285] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1179.852255] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a7982d-fd2e-49dc-a171-d06412ad64f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.861819] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1179.862101] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17e8839a-186f-4f9f-b74f-d1bde197b065 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.923636] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1179.923929] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1179.924201] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleting the datastore file [datastore2] c6d20166-7251-406f-9003-b0599c2f5987 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1179.924496] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4b36a5f-891b-42c5-9edd-1ffe4ef16b14 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.931044] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1179.931044] env[61857]: value = "task-951702" [ 1179.931044] env[61857]: _type = "Task" [ 1179.931044] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.938926] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.131659] env[61857]: DEBUG oslo_vmware.api [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951700, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.441449] env[61857]: DEBUG oslo_vmware.api [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951702, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15924} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.441718] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1180.441848] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1180.442041] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1180.466496] env[61857]: INFO nova.scheduler.client.report [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleted allocations for instance c6d20166-7251-406f-9003-b0599c2f5987 [ 1180.631783] env[61857]: DEBUG oslo_vmware.api [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951700, 'name': PowerOnVM_Task, 'duration_secs': 0.511144} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.632208] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Resumed the VM {{(pid=61857) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1180.632278] env[61857]: DEBUG nova.compute.manager [None req-f4668bd4-6c5a-4e49-b33f-1ef306388110 tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1180.633014] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bd4dc2-e146-4538-8e54-2b4507916218 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.971346] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.971758] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.972095] env[61857]: DEBUG nova.objects.instance [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lazy-loading 'resources' on Instance uuid c6d20166-7251-406f-9003-b0599c2f5987 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1181.475099] env[61857]: DEBUG nova.objects.instance [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lazy-loading 'numa_topology' on Instance uuid c6d20166-7251-406f-9003-b0599c2f5987 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1181.767892] env[61857]: DEBUG nova.compute.manager [req-dea52a16-a0d8-4eb7-8147-7fb2763b8dc9 req-54323e54-7efd-424b-964d-10bd542049d5 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received event network-changed-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1181.768159] env[61857]: DEBUG nova.compute.manager [req-dea52a16-a0d8-4eb7-8147-7fb2763b8dc9 req-54323e54-7efd-424b-964d-10bd542049d5 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Refreshing instance network info cache due to event network-changed-9a0b5327-2d9d-47ef-bb93-6900b47fae64. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1181.768341] env[61857]: DEBUG oslo_concurrency.lockutils [req-dea52a16-a0d8-4eb7-8147-7fb2763b8dc9 req-54323e54-7efd-424b-964d-10bd542049d5 service nova] Acquiring lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.768517] env[61857]: DEBUG oslo_concurrency.lockutils [req-dea52a16-a0d8-4eb7-8147-7fb2763b8dc9 req-54323e54-7efd-424b-964d-10bd542049d5 service nova] Acquired lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.768687] env[61857]: DEBUG nova.network.neutron [req-dea52a16-a0d8-4eb7-8147-7fb2763b8dc9 req-54323e54-7efd-424b-964d-10bd542049d5 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Refreshing network info cache for port 9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1181.978031] env[61857]: DEBUG nova.objects.base [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1182.020098] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f20a3f-8cda-48a3-8a2a-7b8caee17486 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.027522] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8e09c5-238a-4bc0-936a-099b64efb52d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.058833] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92667c54-caba-478d-b658-3095cf30f117 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.065839] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f86ac79-5770-425d-b5f0-e70e9a87b8bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.078410] env[61857]: DEBUG nova.compute.provider_tree [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1182.119766] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "b33cfca7-5621-47a8-bb84-59d82847d699" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.120145] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "b33cfca7-5621-47a8-bb84-59d82847d699" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.120457] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "b33cfca7-5621-47a8-bb84-59d82847d699-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.120747] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "b33cfca7-5621-47a8-bb84-59d82847d699-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.121548] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "b33cfca7-5621-47a8-bb84-59d82847d699-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.124374] env[61857]: INFO nova.compute.manager [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Terminating instance [ 1182.127131] env[61857]: DEBUG nova.compute.manager [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1182.127131] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1182.128254] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370d7782-6d01-4ae6-bf24-2425bfab2da9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.136017] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1182.136444] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58272baa-06b8-4662-856f-84b20e473617 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.142912] env[61857]: DEBUG oslo_vmware.api [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1182.142912] env[61857]: value = "task-951703" [ 1182.142912] env[61857]: _type = "Task" [ 1182.142912] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.152601] env[61857]: DEBUG oslo_vmware.api [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.483413] env[61857]: DEBUG nova.network.neutron [req-dea52a16-a0d8-4eb7-8147-7fb2763b8dc9 req-54323e54-7efd-424b-964d-10bd542049d5 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updated VIF entry in instance network info cache for port 9a0b5327-2d9d-47ef-bb93-6900b47fae64. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1182.483780] env[61857]: DEBUG nova.network.neutron [req-dea52a16-a0d8-4eb7-8147-7fb2763b8dc9 req-54323e54-7efd-424b-964d-10bd542049d5 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updating instance_info_cache with network_info: [{"id": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "address": "fa:16:3e:91:72:71", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9a0b5327-2d", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.581846] env[61857]: DEBUG nova.scheduler.client.report [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1182.652416] env[61857]: DEBUG oslo_vmware.api [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951703, 'name': PowerOffVM_Task, 'duration_secs': 0.187853} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.652683] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1182.652859] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1182.653129] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9da76677-6a51-42f1-84ac-040fe7cee9da {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.714035] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1182.714247] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1182.714426] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleting the datastore file [datastore1] b33cfca7-5621-47a8-bb84-59d82847d699 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1182.714682] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e9b0b8d-2cab-40c6-ae6a-7c1fbedbbff5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.720145] env[61857]: DEBUG oslo_vmware.api [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for the task: (returnval){ [ 1182.720145] env[61857]: value = "task-951705" [ 1182.720145] env[61857]: _type = "Task" [ 1182.720145] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.727878] env[61857]: DEBUG oslo_vmware.api [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951705, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.987577] env[61857]: DEBUG oslo_concurrency.lockutils [req-dea52a16-a0d8-4eb7-8147-7fb2763b8dc9 req-54323e54-7efd-424b-964d-10bd542049d5 service nova] Releasing lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.066230] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "c6d20166-7251-406f-9003-b0599c2f5987" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.086876] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.115s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.231033] env[61857]: DEBUG oslo_vmware.api [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Task: {'id': task-951705, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162524} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.231313] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1183.231504] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1183.231692] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1183.231871] env[61857]: INFO nova.compute.manager [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1183.232134] env[61857]: DEBUG oslo.service.loopingcall [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1183.232336] env[61857]: DEBUG nova.compute.manager [-] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1183.232435] env[61857]: DEBUG nova.network.neutron [-] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1183.597062] env[61857]: DEBUG oslo_concurrency.lockutils [None req-3e0af181-0a73-4b8a-b409-b5b1b66ae51c tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "c6d20166-7251-406f-9003-b0599c2f5987" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.334s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.598741] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "c6d20166-7251-406f-9003-b0599c2f5987" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.533s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.599010] env[61857]: INFO nova.compute.manager [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Unshelving [ 1183.795601] env[61857]: DEBUG nova.compute.manager [req-a9cdfd1f-d3a3-4d50-be80-f6b31211ebec req-7552d828-f60e-4abe-b8a8-894156de87ff service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Received event network-vif-deleted-1c8b2449-1a4b-4062-a760-d88145f9615e {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1183.795795] env[61857]: INFO nova.compute.manager [req-a9cdfd1f-d3a3-4d50-be80-f6b31211ebec req-7552d828-f60e-4abe-b8a8-894156de87ff service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Neutron deleted interface 1c8b2449-1a4b-4062-a760-d88145f9615e; detaching it from the instance and deleting it from the info cache [ 1183.795973] env[61857]: DEBUG nova.network.neutron [req-a9cdfd1f-d3a3-4d50-be80-f6b31211ebec req-7552d828-f60e-4abe-b8a8-894156de87ff service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.116238] env[61857]: DEBUG nova.network.neutron [-] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.299067] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-62f19f42-3d02-42ea-bbee-42421bcf0b77 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.308427] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4143f4f9-7684-48e8-a345-8de61f5258bf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.332449] env[61857]: DEBUG nova.compute.manager [req-a9cdfd1f-d3a3-4d50-be80-f6b31211ebec req-7552d828-f60e-4abe-b8a8-894156de87ff service nova] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Detach interface failed, port_id=1c8b2449-1a4b-4062-a760-d88145f9615e, reason: Instance b33cfca7-5621-47a8-bb84-59d82847d699 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1184.618552] env[61857]: INFO nova.compute.manager [-] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Took 1.39 seconds to deallocate network for instance. [ 1184.624856] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.625136] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.625346] env[61857]: DEBUG nova.objects.instance [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lazy-loading 'pci_requests' on Instance uuid c6d20166-7251-406f-9003-b0599c2f5987 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1185.125397] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.128216] env[61857]: DEBUG nova.objects.instance [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lazy-loading 'numa_topology' on Instance uuid c6d20166-7251-406f-9003-b0599c2f5987 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1185.631343] env[61857]: INFO nova.compute.claims [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1186.685382] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841b8593-b5e4-4d54-8225-ddf364b17bac {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.692666] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22be4fa0-eb46-4d21-8182-1dbb254ed537 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.722283] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59abf33c-65e2-428b-ad56-68a6f41b08da {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.729075] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a729ed2b-7ea0-4374-96de-e6ae7548d667 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.743289] env[61857]: DEBUG nova.compute.provider_tree [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1187.245985] env[61857]: DEBUG nova.scheduler.client.report [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1187.750939] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.126s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.753202] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.628s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.753452] env[61857]: DEBUG nova.objects.instance [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lazy-loading 'resources' on Instance uuid b33cfca7-5621-47a8-bb84-59d82847d699 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.782635] env[61857]: INFO nova.network.neutron [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updating port 9a0b5327-2d9d-47ef-bb93-6900b47fae64 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1188.302325] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef08f13-1dee-4881-b566-87e34a63d7f2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.309824] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ede0d0-b159-40f4-a598-7ebe95a95b2e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.338572] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e514477-21c1-4a94-b928-96740181d0c6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.345664] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fba831f-c303-4797-83bb-da216c165598 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.358785] env[61857]: DEBUG nova.compute.provider_tree [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.862370] env[61857]: DEBUG nova.scheduler.client.report [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1189.368032] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.614s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.385275] env[61857]: INFO nova.scheduler.client.report [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Deleted allocations for instance b33cfca7-5621-47a8-bb84-59d82847d699 [ 1189.841373] env[61857]: DEBUG nova.compute.manager [req-0c32475f-6c73-4fc6-9108-58c3399ff713 req-75e7d259-916f-4e3a-b51b-b41bdd3d67a9 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received event network-vif-plugged-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1189.841608] env[61857]: DEBUG oslo_concurrency.lockutils [req-0c32475f-6c73-4fc6-9108-58c3399ff713 req-75e7d259-916f-4e3a-b51b-b41bdd3d67a9 service nova] Acquiring lock "c6d20166-7251-406f-9003-b0599c2f5987-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.841960] env[61857]: DEBUG oslo_concurrency.lockutils [req-0c32475f-6c73-4fc6-9108-58c3399ff713 req-75e7d259-916f-4e3a-b51b-b41bdd3d67a9 service nova] Lock "c6d20166-7251-406f-9003-b0599c2f5987-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.842116] env[61857]: DEBUG oslo_concurrency.lockutils [req-0c32475f-6c73-4fc6-9108-58c3399ff713 req-75e7d259-916f-4e3a-b51b-b41bdd3d67a9 service nova] Lock "c6d20166-7251-406f-9003-b0599c2f5987-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.842346] env[61857]: DEBUG nova.compute.manager [req-0c32475f-6c73-4fc6-9108-58c3399ff713 req-75e7d259-916f-4e3a-b51b-b41bdd3d67a9 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] No waiting events found dispatching network-vif-plugged-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1189.842481] env[61857]: WARNING nova.compute.manager [req-0c32475f-6c73-4fc6-9108-58c3399ff713 req-75e7d259-916f-4e3a-b51b-b41bdd3d67a9 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received unexpected event network-vif-plugged-9a0b5327-2d9d-47ef-bb93-6900b47fae64 for instance with vm_state shelved_offloaded and task_state spawning. [ 1189.893967] env[61857]: DEBUG oslo_concurrency.lockutils [None req-ed5e5bf5-6008-4a9b-a66b-deb340d7d6fe tempest-ServerActionsTestJSON-1450872981 tempest-ServerActionsTestJSON-1450872981-project-member] Lock "b33cfca7-5621-47a8-bb84-59d82847d699" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.774s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.924006] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.924217] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.924431] env[61857]: DEBUG nova.network.neutron [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1190.631338] env[61857]: DEBUG nova.network.neutron [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updating instance_info_cache with network_info: [{"id": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "address": "fa:16:3e:91:72:71", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a0b5327-2d", "ovs_interfaceid": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.133856] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.160512] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f25930daef56523631ca82fe2d2bfad8',container_format='bare',created_at=2024-09-18T21:48:45Z,direct_url=,disk_format='vmdk',id=16cdeabe-6880-4e6e-9892-9af0bc97d9b3,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1846451837-shelved',owner='35ae3038a282412e963845e221e87f67',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2024-09-18T21:49:00Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1191.160815] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1191.160983] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1191.161198] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1191.161357] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1191.161541] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1191.161755] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1191.161921] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1191.162116] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1191.162292] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1191.162473] env[61857]: DEBUG nova.virt.hardware [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1191.163645] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5aad949-c18b-4832-a411-a89378a580ee {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.171591] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67caa670-9097-45aa-a2ab-c92044342479 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.184627] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:72:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a0b5327-2d9d-47ef-bb93-6900b47fae64', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1191.191929] env[61857]: DEBUG oslo.service.loopingcall [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1191.192174] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1191.192376] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fcdbbcce-5a70-4aca-9ef4-716553567c97 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.211255] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1191.211255] env[61857]: value = "task-951706" [ 1191.211255] env[61857]: _type = "Task" [ 1191.211255] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.218359] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951706, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.720991] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951706, 'name': CreateVM_Task, 'duration_secs': 0.314721} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.721182] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1191.721865] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.722045] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "[datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.722431] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1191.722680] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4541465-0152-4c05-ae90-99031081eb79 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.727039] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1191.727039] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5256df55-bbe1-e412-c1b0-f8e1c2123009" [ 1191.727039] env[61857]: _type = "Task" [ 1191.727039] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.734402] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5256df55-bbe1-e412-c1b0-f8e1c2123009, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.882416] env[61857]: DEBUG nova.compute.manager [req-a5ab23db-57f4-4732-8ab6-2cc2c5bca010 req-5414e834-fcaf-4c44-a2e5-ecc284a07cbe service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received event network-changed-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1191.882416] env[61857]: DEBUG nova.compute.manager [req-a5ab23db-57f4-4732-8ab6-2cc2c5bca010 req-5414e834-fcaf-4c44-a2e5-ecc284a07cbe service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Refreshing instance network info cache due to event network-changed-9a0b5327-2d9d-47ef-bb93-6900b47fae64. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1191.883112] env[61857]: DEBUG oslo_concurrency.lockutils [req-a5ab23db-57f4-4732-8ab6-2cc2c5bca010 req-5414e834-fcaf-4c44-a2e5-ecc284a07cbe service nova] Acquiring lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.883112] env[61857]: DEBUG oslo_concurrency.lockutils [req-a5ab23db-57f4-4732-8ab6-2cc2c5bca010 req-5414e834-fcaf-4c44-a2e5-ecc284a07cbe service nova] Acquired lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.883112] env[61857]: DEBUG nova.network.neutron [req-a5ab23db-57f4-4732-8ab6-2cc2c5bca010 req-5414e834-fcaf-4c44-a2e5-ecc284a07cbe service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Refreshing network info cache for port 9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1192.237730] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "[datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.238095] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Processing image 16cdeabe-6880-4e6e-9892-9af0bc97d9b3 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1192.238262] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3/16cdeabe-6880-4e6e-9892-9af0bc97d9b3.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.238420] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquired lock "[datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3/16cdeabe-6880-4e6e-9892-9af0bc97d9b3.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.238606] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1192.238875] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55fd1578-b95b-4a14-9d09-2c2f776d4d12 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.249154] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1192.249348] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1192.250092] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0b04948-6de9-4ba4-8469-4ba11bc35654 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.255097] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1192.255097] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52b97837-4454-7a69-4489-bc9a4464f2e8" [ 1192.255097] env[61857]: _type = "Task" [ 1192.255097] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.262755] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52b97837-4454-7a69-4489-bc9a4464f2e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.602826] env[61857]: DEBUG nova.network.neutron [req-a5ab23db-57f4-4732-8ab6-2cc2c5bca010 req-5414e834-fcaf-4c44-a2e5-ecc284a07cbe service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updated VIF entry in instance network info cache for port 9a0b5327-2d9d-47ef-bb93-6900b47fae64. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1192.603240] env[61857]: DEBUG nova.network.neutron [req-a5ab23db-57f4-4732-8ab6-2cc2c5bca010 req-5414e834-fcaf-4c44-a2e5-ecc284a07cbe service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updating instance_info_cache with network_info: [{"id": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "address": "fa:16:3e:91:72:71", "network": {"id": "b6300051-3996-4479-b289-d683b156f319", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1649504940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35ae3038a282412e963845e221e87f67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a0b5327-2d", "ovs_interfaceid": "9a0b5327-2d9d-47ef-bb93-6900b47fae64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.765858] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Preparing fetch location {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1192.766075] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Fetch image to [datastore1] OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f/OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f.vmdk {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1192.766269] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Downloading stream optimized image 16cdeabe-6880-4e6e-9892-9af0bc97d9b3 to [datastore1] OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f/OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f.vmdk on the data store datastore1 as vApp {{(pid=61857) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1192.766448] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Downloading image file data 16cdeabe-6880-4e6e-9892-9af0bc97d9b3 to the ESX as VM named 'OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f' {{(pid=61857) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1192.837632] env[61857]: DEBUG oslo_vmware.rw_handles [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1192.837632] env[61857]: value = "resgroup-9" [ 1192.837632] env[61857]: _type = "ResourcePool" [ 1192.837632] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1192.837901] env[61857]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-31fc5a63-3e4a-449c-807a-ad56c1fb91eb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.858223] env[61857]: DEBUG oslo_vmware.rw_handles [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lease: (returnval){ [ 1192.858223] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e26a7e-e996-c733-93f7-48aebce18d53" [ 1192.858223] env[61857]: _type = "HttpNfcLease" [ 1192.858223] env[61857]: } obtained for vApp import into resource pool (val){ [ 1192.858223] env[61857]: value = "resgroup-9" [ 1192.858223] env[61857]: _type = "ResourcePool" [ 1192.858223] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1192.858543] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the lease: (returnval){ [ 1192.858543] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e26a7e-e996-c733-93f7-48aebce18d53" [ 1192.858543] env[61857]: _type = "HttpNfcLease" [ 1192.858543] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1192.864446] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1192.864446] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e26a7e-e996-c733-93f7-48aebce18d53" [ 1192.864446] env[61857]: _type = "HttpNfcLease" [ 1192.864446] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1193.106542] env[61857]: DEBUG oslo_concurrency.lockutils [req-a5ab23db-57f4-4732-8ab6-2cc2c5bca010 req-5414e834-fcaf-4c44-a2e5-ecc284a07cbe service nova] Releasing lock "refresh_cache-c6d20166-7251-406f-9003-b0599c2f5987" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.365930] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1193.365930] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e26a7e-e996-c733-93f7-48aebce18d53" [ 1193.365930] env[61857]: _type = "HttpNfcLease" [ 1193.365930] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1193.866768] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1193.866768] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e26a7e-e996-c733-93f7-48aebce18d53" [ 1193.866768] env[61857]: _type = "HttpNfcLease" [ 1193.866768] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1193.867084] env[61857]: DEBUG oslo_vmware.rw_handles [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1193.867084] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52e26a7e-e996-c733-93f7-48aebce18d53" [ 1193.867084] env[61857]: _type = "HttpNfcLease" [ 1193.867084] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1193.867774] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f7ea3b-db09-4df9-be38-f8d1639ed450 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.874954] env[61857]: DEBUG oslo_vmware.rw_handles [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52002e3c-5ac2-d032-17c9-269eb83277d0/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1193.875153] env[61857]: DEBUG oslo_vmware.rw_handles [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52002e3c-5ac2-d032-17c9-269eb83277d0/disk-0.vmdk. {{(pid=61857) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1193.937156] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-98b424d7-1f81-40d7-adee-5823a570bacc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.076196] env[61857]: DEBUG oslo_vmware.rw_handles [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Completed reading data from the image iterator. {{(pid=61857) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1195.076606] env[61857]: DEBUG oslo_vmware.rw_handles [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52002e3c-5ac2-d032-17c9-269eb83277d0/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1195.077436] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93759668-79d2-4e78-a743-bd6aed4626ef {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.084136] env[61857]: DEBUG oslo_vmware.rw_handles [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52002e3c-5ac2-d032-17c9-269eb83277d0/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1195.084376] env[61857]: DEBUG oslo_vmware.rw_handles [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52002e3c-5ac2-d032-17c9-269eb83277d0/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1195.084530] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9208734e-7c14-4480-9a05-8dcbec000184 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.256371] env[61857]: DEBUG oslo_vmware.rw_handles [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52002e3c-5ac2-d032-17c9-269eb83277d0/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1195.256588] env[61857]: INFO nova.virt.vmwareapi.images [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Downloaded image file data 16cdeabe-6880-4e6e-9892-9af0bc97d9b3 [ 1195.257411] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e254a7-cae2-49ef-bb2a-bf7850066584 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.272574] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7376278c-5766-4ef9-a1c2-06d444b22ff3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.308226] env[61857]: INFO nova.virt.vmwareapi.images [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] The imported VM was unregistered [ 1195.310564] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Caching image {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1195.310800] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Creating directory with path [datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3 {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1195.311071] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1ee9525-95a8-49a3-b5bf-b868dd77d4de {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.320389] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Created directory with path [datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3 {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1195.320584] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f/OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f.vmdk to [datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3/16cdeabe-6880-4e6e-9892-9af0bc97d9b3.vmdk. {{(pid=61857) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1195.320814] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-5fd451f5-dbb7-49b6-aab7-e08f2c1c148f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.326168] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1195.326168] env[61857]: value = "task-951709" [ 1195.326168] env[61857]: _type = "Task" [ 1195.326168] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.333146] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951709, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.836489] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951709, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.336653] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951709, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.837473] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951709, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.339493] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951709, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.838400] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951709, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.134453} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.838660] env[61857]: INFO nova.virt.vmwareapi.ds_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f/OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f.vmdk to [datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3/16cdeabe-6880-4e6e-9892-9af0bc97d9b3.vmdk. [ 1197.838851] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Cleaning up location [datastore1] OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1197.839030] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_4fe8139d-6413-46c8-8c61-c7923a3bd26f {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1197.839287] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-69a6686a-063b-49d5-adda-2c6498ba51df {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.845017] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1197.845017] env[61857]: value = "task-951710" [ 1197.845017] env[61857]: _type = "Task" [ 1197.845017] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.851924] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951710, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.355241] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951710, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.040151} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.355638] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1198.355677] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Releasing lock "[datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3/16cdeabe-6880-4e6e-9892-9af0bc97d9b3.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.355910] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3/16cdeabe-6880-4e6e-9892-9af0bc97d9b3.vmdk to [datastore1] c6d20166-7251-406f-9003-b0599c2f5987/c6d20166-7251-406f-9003-b0599c2f5987.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1198.356197] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17b85825-59bf-4da0-99dc-eec9447e3218 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.363442] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1198.363442] env[61857]: value = "task-951711" [ 1198.363442] env[61857]: _type = "Task" [ 1198.363442] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.372429] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951711, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.873753] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951711, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.375349] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951711, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.876274] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951711, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.377773] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951711, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.690218] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.877223] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951711, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.107884} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.877518] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/16cdeabe-6880-4e6e-9892-9af0bc97d9b3/16cdeabe-6880-4e6e-9892-9af0bc97d9b3.vmdk to [datastore1] c6d20166-7251-406f-9003-b0599c2f5987/c6d20166-7251-406f-9003-b0599c2f5987.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1200.878305] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b5c485-5deb-486b-8bfd-bb6fd9eb224b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.899795] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] c6d20166-7251-406f-9003-b0599c2f5987/c6d20166-7251-406f-9003-b0599c2f5987.vmdk or device None with type streamOptimized {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1200.900031] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f9a0779-2041-4155-bdc1-53c175ca3c90 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.918870] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1200.918870] env[61857]: value = "task-951712" [ 1200.918870] env[61857]: _type = "Task" [ 1200.918870] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.926213] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951712, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.428296] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951712, 'name': ReconfigVM_Task, 'duration_secs': 0.288064} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.428669] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Reconfigured VM instance instance-0000006a to attach disk [datastore1] c6d20166-7251-406f-9003-b0599c2f5987/c6d20166-7251-406f-9003-b0599c2f5987.vmdk or device None with type streamOptimized {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1201.429227] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74c0128a-dce6-415c-ad8e-51a0117f500a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.435691] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1201.435691] env[61857]: value = "task-951713" [ 1201.435691] env[61857]: _type = "Task" [ 1201.435691] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.447831] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951713, 'name': Rename_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.945467] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951713, 'name': Rename_Task, 'duration_secs': 0.134551} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.945759] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1201.946000] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8892043-ff77-4639-98fd-267f6f5cbebf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.951671] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1201.951671] env[61857]: value = "task-951714" [ 1201.951671] env[61857]: _type = "Task" [ 1201.951671] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.958748] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951714, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.463557] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951714, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.689532] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.962028] env[61857]: DEBUG oslo_vmware.api [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951714, 'name': PowerOnVM_Task, 'duration_secs': 0.543488} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.962323] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1203.064175] env[61857]: DEBUG nova.compute.manager [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1203.065108] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c5a6d7-483f-4cef-b026-a670341c047d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.582213] env[61857]: DEBUG oslo_concurrency.lockutils [None req-b9d4a75d-2fab-4856-8209-afd742a421e2 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "c6d20166-7251-406f-9003-b0599c2f5987" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.983s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.685179] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.305754] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "299da51d-22a9-4da5-992a-482b63b81034" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.306026] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.689622] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.689859] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1205.809014] env[61857]: DEBUG nova.compute.utils [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1206.312090] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.372680] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "299da51d-22a9-4da5-992a-482b63b81034" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.372680] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.373050] env[61857]: INFO nova.compute.manager [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Attaching volume 454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb to /dev/sdb [ 1207.405599] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13cd811-6148-4137-a92e-a49422a1dbeb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.412740] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27e3bd9-93de-4ad6-946f-c1d6a6b18916 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.426214] env[61857]: DEBUG nova.virt.block_device [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating existing volume attachment record: 0e7cbe12-a9a2-41a6-bd30-ea554298027f {{(pid=61857) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1207.699631] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Didn't find any instances for network info cache update. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1207.699886] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1207.700071] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1207.700260] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1207.700487] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1208.203524] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.203703] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.203859] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.204022] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1208.204916] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e9aa47-76fa-404f-a13f-d1f7e21aaa91 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.213118] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eea9c2a-5556-449a-b3a6-21df9e4d188a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.226456] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70742720-a8e8-420c-8dbd-1945bdca9374 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.232544] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437c0b60-a467-450c-b785-c37467fee76b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.260551] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181080MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1208.260686] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.260874] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.288006] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance 299da51d-22a9-4da5-992a-482b63b81034 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1209.288374] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c6d20166-7251-406f-9003-b0599c2f5987 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1209.288374] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1209.288523] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1209.325237] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3200e75-03dd-454e-baf7-530e80414481 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.332899] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdff9755-6ce9-4f4e-927f-d9edf42c984e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.365626] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52f6098-96a3-405b-9b8a-cffd8bf5c781 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.373150] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a658a66-b79a-40e5-8e6c-1a654ad1ed5f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.386395] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.889216] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1210.394452] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1210.394929] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.134s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1211.384605] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1211.889712] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1211.890013] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1211.969680] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Volume attach. Driver type: vmdk {{(pid=61857) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1211.969945] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214237', 'volume_id': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'name': 'volume-454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '299da51d-22a9-4da5-992a-482b63b81034', 'attached_at': '', 'detached_at': '', 'volume_id': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'serial': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1211.970898] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7d2bc5-50e9-40a0-93b4-17a18e11362d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.987141] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b64c6f-7c83-4147-ae5b-71afeb669f43 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.012738] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb/volume-454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1212.013068] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db906a21-d34d-4117-98d1-2b10a52818c9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.031348] env[61857]: DEBUG oslo_vmware.api [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1212.031348] env[61857]: value = "task-951719" [ 1212.031348] env[61857]: _type = "Task" [ 1212.031348] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.041396] env[61857]: DEBUG oslo_vmware.api [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951719, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.541073] env[61857]: DEBUG oslo_vmware.api [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951719, 'name': ReconfigVM_Task, 'duration_secs': 0.335453} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.541454] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb/volume-454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1212.546130] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a5c095b-b5b9-4319-8a28-d72ea37f0dbd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.560549] env[61857]: DEBUG oslo_vmware.api [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1212.560549] env[61857]: value = "task-951720" [ 1212.560549] env[61857]: _type = "Task" [ 1212.560549] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.567965] env[61857]: DEBUG oslo_vmware.api [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951720, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.070293] env[61857]: DEBUG oslo_vmware.api [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951720, 'name': ReconfigVM_Task, 'duration_secs': 0.129989} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.070667] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214237', 'volume_id': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'name': 'volume-454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '299da51d-22a9-4da5-992a-482b63b81034', 'attached_at': '', 'detached_at': '', 'volume_id': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'serial': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb'} {{(pid=61857) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1214.107854] env[61857]: DEBUG nova.objects.instance [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'flavor' on Instance uuid 299da51d-22a9-4da5-992a-482b63b81034 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.613382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-a7f5507c-0d62-4713-8778-5ad3adde58dd tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.240s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.427274] env[61857]: DEBUG nova.compute.manager [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Stashing vm_state: active {{(pid=61857) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1215.947151] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.947447] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.452155] env[61857]: INFO nova.compute.claims [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1216.958219] env[61857]: INFO nova.compute.resource_tracker [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating resource usage from migration fe21ee12-24bb-44ba-a21b-71e26ed4040a [ 1217.005135] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b5882a-e033-48ff-97a6-37269336b90d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.012791] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4527c76b-21f7-49b5-9304-02c062338205 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.042688] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985e69b6-3b88-4ec1-8b5e-d5e02269bfb1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.049508] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559d0e3b-0ddc-4908-9b97-6680be0101e2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.062034] env[61857]: DEBUG nova.compute.provider_tree [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.565752] env[61857]: DEBUG nova.scheduler.client.report [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1218.071382] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.124s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.071677] env[61857]: INFO nova.compute.manager [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Migrating [ 1218.586928] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.587327] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.587371] env[61857]: DEBUG nova.network.neutron [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1219.289571] env[61857]: DEBUG nova.network.neutron [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance_info_cache with network_info: [{"id": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "address": "fa:16:3e:0b:95:f2", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6a429a1-74", "ovs_interfaceid": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.792503] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.307470] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4182b26-0f0e-4b8f-aa62-17cf2781d31e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.328432] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance '299da51d-22a9-4da5-992a-482b63b81034' progress to 0 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1221.834631] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1221.834957] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5729c7a7-b10e-4b73-a98a-68a19b526270 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.842021] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1221.842021] env[61857]: value = "task-951721" [ 1221.842021] env[61857]: _type = "Task" [ 1221.842021] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.352135] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951721, 'name': PowerOffVM_Task, 'duration_secs': 0.247762} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.352468] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1222.352590] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance '299da51d-22a9-4da5-992a-482b63b81034' progress to 17 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1222.859024] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1222.859327] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1222.859474] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1222.859667] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1222.859822] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1222.859976] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1222.860201] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1222.860395] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1222.860568] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1222.860739] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1222.860919] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1222.865899] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42ba11ae-94d7-4dee-9b68-59fd9d71ede0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.882056] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1222.882056] env[61857]: value = "task-951722" [ 1222.882056] env[61857]: _type = "Task" [ 1222.882056] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.889710] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951722, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.391734] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951722, 'name': ReconfigVM_Task, 'duration_secs': 0.17155} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.392119] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance '299da51d-22a9-4da5-992a-482b63b81034' progress to 33 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1223.898933] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1223.899249] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1223.899469] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1223.899694] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1223.899889] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1223.900093] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1223.900360] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1223.900576] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1223.900788] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1223.900995] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1223.901239] env[61857]: DEBUG nova.virt.hardware [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1223.906506] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1223.906813] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83e31745-7def-4521-91b3-f98bc940b0db {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.925908] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1223.925908] env[61857]: value = "task-951723" [ 1223.925908] env[61857]: _type = "Task" [ 1223.925908] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.933208] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951723, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.435226] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951723, 'name': ReconfigVM_Task, 'duration_secs': 0.170116} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.435553] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1224.436283] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0e6699-7c00-4c66-9d8c-ec778f670754 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.460382] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 299da51d-22a9-4da5-992a-482b63b81034/299da51d-22a9-4da5-992a-482b63b81034.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1224.460973] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ced03c6-50ce-41c5-a847-91ddbe9b10c0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.478595] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1224.478595] env[61857]: value = "task-951724" [ 1224.478595] env[61857]: _type = "Task" [ 1224.478595] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.486322] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951724, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.989090] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951724, 'name': ReconfigVM_Task, 'duration_secs': 0.288571} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.989390] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 299da51d-22a9-4da5-992a-482b63b81034/299da51d-22a9-4da5-992a-482b63b81034.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1224.989694] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance '299da51d-22a9-4da5-992a-482b63b81034' progress to 50 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1225.497382] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad27f2d-4f08-4c44-9c10-7ab217654f70 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.518753] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376d9d3a-e1ec-4ab4-94a5-e87e11a73649 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.539999] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance '299da51d-22a9-4da5-992a-482b63b81034' progress to 67 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1227.175812] env[61857]: DEBUG nova.network.neutron [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Port c6a429a1-74cb-4ee8-8024-7d25225c2265 binding to destination host cpu-1 is already ACTIVE {{(pid=61857) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1228.197239] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "299da51d-22a9-4da5-992a-482b63b81034-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1228.197616] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1228.197616] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.231196] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1229.231531] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1229.231640] env[61857]: DEBUG nova.network.neutron [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1229.932715] env[61857]: DEBUG nova.network.neutron [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance_info_cache with network_info: [{"id": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "address": "fa:16:3e:0b:95:f2", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6a429a1-74", "ovs_interfaceid": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.435766] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1230.945309] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c308722a-f224-4050-955f-f5a8b855f1aa {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.952447] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95644bb7-4eff-4eb8-a24a-dfe86a79053e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.044887] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c2b2b1-cd1a-45ef-8c4f-2cddc2cf8232 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.066771] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3c11cb-e0a3-4d4c-a03e-60bafebdca15 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.073212] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance '299da51d-22a9-4da5-992a-482b63b81034' progress to 83 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1232.579899] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1232.580204] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14936a1b-5dea-4979-8e67-6cee1c6e0ecc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.588272] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1232.588272] env[61857]: value = "task-951725" [ 1232.588272] env[61857]: _type = "Task" [ 1232.588272] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.596057] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.097899] env[61857]: DEBUG oslo_vmware.api [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951725, 'name': PowerOnVM_Task, 'duration_secs': 0.410651} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.098365] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1233.098466] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9def5b3c-d8f2-4b29-8e26-c1ad38afc0de tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance '299da51d-22a9-4da5-992a-482b63b81034' progress to 100 {{(pid=61857) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1235.336442] env[61857]: DEBUG nova.network.neutron [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Port c6a429a1-74cb-4ee8-8024-7d25225c2265 binding to destination host cpu-1 is already ACTIVE {{(pid=61857) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1235.336717] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1235.336876] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.337059] env[61857]: DEBUG nova.network.neutron [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1236.111549] env[61857]: DEBUG nova.network.neutron [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance_info_cache with network_info: [{"id": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "address": "fa:16:3e:0b:95:f2", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6a429a1-74", "ovs_interfaceid": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1236.614116] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1237.117924] env[61857]: DEBUG nova.compute.manager [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61857) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1238.210287] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.210645] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.713781] env[61857]: DEBUG nova.objects.instance [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'migration_context' on Instance uuid 299da51d-22a9-4da5-992a-482b63b81034 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1239.270243] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6296551e-e16f-4f2e-8cdb-8c2a45044d97 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.278009] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7091b9-81d4-4e83-8872-5dda72603c28 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.308424] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff94edb3-22c5-433f-8bf6-dbb97c2c501d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.315364] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa0dda2-32d3-451b-8b79-68a2a1e6b0d9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.327878] env[61857]: DEBUG nova.compute.provider_tree [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1239.831215] env[61857]: DEBUG nova.scheduler.client.report [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1240.290562] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "c6d20166-7251-406f-9003-b0599c2f5987" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.290940] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "c6d20166-7251-406f-9003-b0599c2f5987" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.291083] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "c6d20166-7251-406f-9003-b0599c2f5987-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.291278] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "c6d20166-7251-406f-9003-b0599c2f5987-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.291455] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "c6d20166-7251-406f-9003-b0599c2f5987-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.293660] env[61857]: INFO nova.compute.manager [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Terminating instance [ 1240.295394] env[61857]: DEBUG nova.compute.manager [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1240.295592] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1240.296417] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37485592-740b-421b-913d-581d25b5413a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.304159] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1240.304884] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad3dc88b-5e5d-487d-9c25-4f43117c8d06 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.310637] env[61857]: DEBUG oslo_vmware.api [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1240.310637] env[61857]: value = "task-951726" [ 1240.310637] env[61857]: _type = "Task" [ 1240.310637] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.317979] env[61857]: DEBUG oslo_vmware.api [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951726, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.820453] env[61857]: DEBUG oslo_vmware.api [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951726, 'name': PowerOffVM_Task, 'duration_secs': 0.168332} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.820742] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1240.820923] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1240.821195] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af015ebd-bb98-4440-a24c-cacc417caab0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.842315] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.632s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.880142] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1240.880399] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Deleting contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1240.880595] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleting the datastore file [datastore1] c6d20166-7251-406f-9003-b0599c2f5987 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1240.880855] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e5a16a9-a660-41a9-826a-73f214a1043a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.888064] env[61857]: DEBUG oslo_vmware.api [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for the task: (returnval){ [ 1240.888064] env[61857]: value = "task-951728" [ 1240.888064] env[61857]: _type = "Task" [ 1240.888064] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.895279] env[61857]: DEBUG oslo_vmware.api [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951728, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.398155] env[61857]: DEBUG oslo_vmware.api [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Task: {'id': task-951728, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.114853} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.398557] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1241.398612] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Deleted contents of the VM from datastore datastore1 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1241.398774] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1241.398952] env[61857]: INFO nova.compute.manager [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1241.399220] env[61857]: DEBUG oslo.service.loopingcall [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1241.399422] env[61857]: DEBUG nova.compute.manager [-] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1241.399519] env[61857]: DEBUG nova.network.neutron [-] [instance: c6d20166-7251-406f-9003-b0599c2f5987] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1241.835279] env[61857]: DEBUG nova.compute.manager [req-f02912d9-f724-46b4-9d1a-a135f7ea73d7 req-4e5d102d-b01a-4d23-ad1f-9fe0fe52f8b8 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Received event network-vif-deleted-9a0b5327-2d9d-47ef-bb93-6900b47fae64 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1241.835388] env[61857]: INFO nova.compute.manager [req-f02912d9-f724-46b4-9d1a-a135f7ea73d7 req-4e5d102d-b01a-4d23-ad1f-9fe0fe52f8b8 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Neutron deleted interface 9a0b5327-2d9d-47ef-bb93-6900b47fae64; detaching it from the instance and deleting it from the info cache [ 1241.835554] env[61857]: DEBUG nova.network.neutron [req-f02912d9-f724-46b4-9d1a-a135f7ea73d7 req-4e5d102d-b01a-4d23-ad1f-9fe0fe52f8b8 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.313833] env[61857]: DEBUG nova.network.neutron [-] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.338449] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f60a56a-1d08-4b91-b4ab-3306f7a727b8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.348424] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c84585-8223-476b-bd6d-5b90c06f6d15 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.372050] env[61857]: DEBUG nova.compute.manager [req-f02912d9-f724-46b4-9d1a-a135f7ea73d7 req-4e5d102d-b01a-4d23-ad1f-9fe0fe52f8b8 service nova] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Detach interface failed, port_id=9a0b5327-2d9d-47ef-bb93-6900b47fae64, reason: Instance c6d20166-7251-406f-9003-b0599c2f5987 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1242.378198] env[61857]: INFO nova.compute.manager [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Swapping old allocation on dict_keys(['2d5860d0-8574-4e55-9ced-20e66f0314c2']) held by migration fe21ee12-24bb-44ba-a21b-71e26ed4040a for instance [ 1242.401260] env[61857]: DEBUG nova.scheduler.client.report [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Overwriting current allocation {'allocations': {'2d5860d0-8574-4e55-9ced-20e66f0314c2': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 149}}, 'project_id': '2a7d2f228f88485f863b3c421b25a37c', 'user_id': 'f4c9ee17921646978e8b21ad5508813c', 'consumer_generation': 1} on consumer 299da51d-22a9-4da5-992a-482b63b81034 {{(pid=61857) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1242.560978] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.561206] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.561397] env[61857]: DEBUG nova.network.neutron [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1242.817229] env[61857]: INFO nova.compute.manager [-] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Took 1.42 seconds to deallocate network for instance. [ 1243.263604] env[61857]: DEBUG nova.network.neutron [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance_info_cache with network_info: [{"id": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "address": "fa:16:3e:0b:95:f2", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6a429a1-74", "ovs_interfaceid": "c6a429a1-74cb-4ee8-8024-7d25225c2265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1243.323381] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.323663] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.323890] env[61857]: DEBUG nova.objects.instance [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lazy-loading 'resources' on Instance uuid c6d20166-7251-406f-9003-b0599c2f5987 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1243.766728] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-299da51d-22a9-4da5-992a-482b63b81034" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1243.767777] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2176f769-7f7d-409c-b356-26f305431780 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.775259] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cd9db6-635e-427d-9aad-05a2ef8cbd00 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.867202] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f5aeb7-36ec-4d5d-abd3-7a621cf49484 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.873706] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093ac87d-604f-43cc-bb25-2eaf7a5a3426 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.902217] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc6b16c-99b2-4aba-988a-20cdc9f598c2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.908731] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e92b140-495c-48bb-9989-b41f75044aab {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.921052] env[61857]: DEBUG nova.compute.provider_tree [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1244.423529] env[61857]: DEBUG nova.scheduler.client.report [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1244.855499] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1244.855809] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ed68dfe-90f7-4410-af84-589896e62c00 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.862900] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1244.862900] env[61857]: value = "task-951729" [ 1244.862900] env[61857]: _type = "Task" [ 1244.862900] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.872015] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.928332] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.604s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.946544] env[61857]: INFO nova.scheduler.client.report [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Deleted allocations for instance c6d20166-7251-406f-9003-b0599c2f5987 [ 1245.373025] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951729, 'name': PowerOffVM_Task, 'duration_secs': 0.192377} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.373025] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1245.373177] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1245.373300] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1245.373471] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1245.373658] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1245.373810] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1245.373960] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1245.374175] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1245.374337] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1245.374556] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1245.374755] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1245.374962] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1245.379869] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d2544fa-0008-4d1e-8543-e6ceb79c6b4d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.394692] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1245.394692] env[61857]: value = "task-951730" [ 1245.394692] env[61857]: _type = "Task" [ 1245.394692] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.402265] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951730, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.454651] env[61857]: DEBUG oslo_concurrency.lockutils [None req-c863ad8e-2876-434a-bd4f-957dfbcb7f26 tempest-AttachVolumeShelveTestJSON-544980617 tempest-AttachVolumeShelveTestJSON-544980617-project-member] Lock "c6d20166-7251-406f-9003-b0599c2f5987" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.164s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.904782] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951730, 'name': ReconfigVM_Task, 'duration_secs': 0.142} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.905569] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6336bca0-ca53-4858-a18d-115c0b7cb7ed {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.926034] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1245.926286] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1245.926420] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1245.926614] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1245.926770] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1245.926925] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1245.927149] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1245.927321] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1245.927499] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1245.927675] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1245.927859] env[61857]: DEBUG nova.virt.hardware [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1245.928701] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-898076ac-8e7f-4895-a315-513079979b05 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.933941] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1245.933941] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52bb2e22-30ef-f74d-3aa1-001ae17aea4b" [ 1245.933941] env[61857]: _type = "Task" [ 1245.933941] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.941546] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52bb2e22-30ef-f74d-3aa1-001ae17aea4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.443753] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52bb2e22-30ef-f74d-3aa1-001ae17aea4b, 'name': SearchDatastore_Task, 'duration_secs': 0.006965} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.449036] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1246.449310] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03d79b84-6226-4f79-9719-4eb7286b7dc3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.466900] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1246.466900] env[61857]: value = "task-951731" [ 1246.466900] env[61857]: _type = "Task" [ 1246.466900] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.474048] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951731, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.977398] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951731, 'name': ReconfigVM_Task, 'duration_secs': 0.193113} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.977785] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1246.978446] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba2068f-5a24-4d20-a123-99f09edd0172 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.002115] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 299da51d-22a9-4da5-992a-482b63b81034/299da51d-22a9-4da5-992a-482b63b81034.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1247.002358] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-262e37c0-28fc-4ef0-aa4a-ac7d29e768f3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.019339] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1247.019339] env[61857]: value = "task-951733" [ 1247.019339] env[61857]: _type = "Task" [ 1247.019339] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.027771] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951733, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.529906] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951733, 'name': ReconfigVM_Task, 'duration_secs': 0.262622} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.530226] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 299da51d-22a9-4da5-992a-482b63b81034/299da51d-22a9-4da5-992a-482b63b81034.vmdk or device None with type thin {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1247.531074] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008c66a1-c8e2-41c3-b447-22edc61ab340 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.551104] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e59306-2afc-4354-9732-93d953bd838a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.570726] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ba9ecd-8d76-4c8d-b475-59a9a0bc2516 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.590124] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fc7fc2-e16d-4235-b1e8-4b451480ddc2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.596602] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1247.596838] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31ad0d3b-a5ff-4508-a573-634dbd5c66d6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.602498] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1247.602498] env[61857]: value = "task-951734" [ 1247.602498] env[61857]: _type = "Task" [ 1247.602498] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.609296] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951734, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.111724] env[61857]: DEBUG oslo_vmware.api [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951734, 'name': PowerOnVM_Task, 'duration_secs': 0.359518} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.112177] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1249.152247] env[61857]: INFO nova.compute.manager [None req-c703fb5e-e5db-4649-ae44-5858fd03b038 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance to original state: 'active' [ 1250.162707] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "299da51d-22a9-4da5-992a-482b63b81034" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.164078] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.164078] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "299da51d-22a9-4da5-992a-482b63b81034-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.164078] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.164078] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1250.167818] env[61857]: INFO nova.compute.manager [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Terminating instance [ 1250.169859] env[61857]: DEBUG nova.compute.manager [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1250.170081] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1250.170371] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-360e785b-2809-4135-b523-91455c87e5b0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.178143] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1250.178143] env[61857]: value = "task-951735" [ 1250.178143] env[61857]: _type = "Task" [ 1250.178143] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.187482] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.693070] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951735, 'name': PowerOffVM_Task, 'duration_secs': 0.191991} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.693070] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1250.693070] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Volume detach. Driver type: vmdk {{(pid=61857) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1250.693070] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214237', 'volume_id': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'name': 'volume-454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '299da51d-22a9-4da5-992a-482b63b81034', 'attached_at': '2024-09-18T21:50:07.000000', 'detached_at': '', 'volume_id': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'serial': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1250.693070] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ce5a27-926d-44fe-9c9e-f35fd032768d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.716587] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf9b93d-5f2b-4313-9222-2a01bad08b07 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.723831] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7dc97c-98f9-41f9-9d15-840a7fc32144 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.746249] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b292a5-90a3-44dc-8a94-c722e3b959ce {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.760883] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] The volume has not been displaced from its original location: [datastore2] volume-454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb/volume-454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb.vmdk. No consolidation needed. {{(pid=61857) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1250.766209] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1250.766493] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6777cfad-4f68-4516-974a-b7b8e0ec8d00 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.785566] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1250.785566] env[61857]: value = "task-951736" [ 1250.785566] env[61857]: _type = "Task" [ 1250.785566] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.795197] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951736, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.296505] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951736, 'name': ReconfigVM_Task, 'duration_secs': 0.182471} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.296844] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=61857) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1251.301714] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-054601c9-54fb-4009-8d4d-32400ee7713e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.316335] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1251.316335] env[61857]: value = "task-951737" [ 1251.316335] env[61857]: _type = "Task" [ 1251.316335] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.324824] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951737, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.826141] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951737, 'name': ReconfigVM_Task, 'duration_secs': 0.126219} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.826472] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-214237', 'volume_id': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'name': 'volume-454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '299da51d-22a9-4da5-992a-482b63b81034', 'attached_at': '2024-09-18T21:50:07.000000', 'detached_at': '', 'volume_id': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb', 'serial': '454a4c53-fdc0-4ef3-9f61-6b3d0b9205fb'} {{(pid=61857) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1251.826769] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1251.827520] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb86806-f396-445a-a3da-b5f5208d57ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.834264] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1251.834489] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6401953a-d4eb-445a-accd-92ec14efb0c0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.899248] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1251.899497] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1251.899709] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleting the datastore file [datastore2] 299da51d-22a9-4da5-992a-482b63b81034 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1251.899972] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eec41f94-491c-4d41-80a2-fdb7ee56efc0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.905594] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1251.905594] env[61857]: value = "task-951739" [ 1251.905594] env[61857]: _type = "Task" [ 1251.905594] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.913208] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951739, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.415498] env[61857]: DEBUG oslo_vmware.api [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951739, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144626} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.417062] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1252.417062] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1252.417062] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1252.417062] env[61857]: INFO nova.compute.manager [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Took 2.25 seconds to destroy the instance on the hypervisor. [ 1252.417062] env[61857]: DEBUG oslo.service.loopingcall [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1252.417319] env[61857]: DEBUG nova.compute.manager [-] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1252.417361] env[61857]: DEBUG nova.network.neutron [-] [instance: 299da51d-22a9-4da5-992a-482b63b81034] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1252.818817] env[61857]: DEBUG nova.compute.manager [req-f8007691-a222-4fb1-a2df-00239d116a0e req-cfc7e3a6-6427-46c3-be74-533713f6d68e service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Received event network-vif-deleted-c6a429a1-74cb-4ee8-8024-7d25225c2265 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1252.819028] env[61857]: INFO nova.compute.manager [req-f8007691-a222-4fb1-a2df-00239d116a0e req-cfc7e3a6-6427-46c3-be74-533713f6d68e service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Neutron deleted interface c6a429a1-74cb-4ee8-8024-7d25225c2265; detaching it from the instance and deleting it from the info cache [ 1252.819173] env[61857]: DEBUG nova.network.neutron [req-f8007691-a222-4fb1-a2df-00239d116a0e req-cfc7e3a6-6427-46c3-be74-533713f6d68e service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1253.295446] env[61857]: DEBUG nova.network.neutron [-] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1253.322563] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6feff405-83c3-4240-9727-60c091d0f2be {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.331883] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc48848-ad73-4453-9f94-2e54ed6a6096 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.354561] env[61857]: DEBUG nova.compute.manager [req-f8007691-a222-4fb1-a2df-00239d116a0e req-cfc7e3a6-6427-46c3-be74-533713f6d68e service nova] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Detach interface failed, port_id=c6a429a1-74cb-4ee8-8024-7d25225c2265, reason: Instance 299da51d-22a9-4da5-992a-482b63b81034 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1253.798440] env[61857]: INFO nova.compute.manager [-] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Took 1.38 seconds to deallocate network for instance. [ 1254.339743] env[61857]: INFO nova.compute.manager [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Took 0.54 seconds to detach 1 volumes for instance. [ 1254.847545] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.847912] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.848051] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.868191] env[61857]: INFO nova.scheduler.client.report [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleted allocations for instance 299da51d-22a9-4da5-992a-482b63b81034 [ 1255.376123] env[61857]: DEBUG oslo_concurrency.lockutils [None req-9d48e525-4772-4f74-a85c-193b2c0e2ec6 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "299da51d-22a9-4da5-992a-482b63b81034" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.213s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.310114] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1256.310434] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1256.813054] env[61857]: DEBUG nova.compute.manager [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1257.334197] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1257.334470] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1257.336309] env[61857]: INFO nova.compute.claims [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1258.370898] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db5db24-2268-4fe4-9c51-214e2115e4b4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.378637] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151a30fa-216d-4303-9a7f-6580615135a2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.408996] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b22040a-f180-4d5d-8eba-706e850862e9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.416601] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8968c658-00d7-4b40-a9c3-d134f3989b19 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.430009] env[61857]: DEBUG nova.compute.provider_tree [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1258.933489] env[61857]: DEBUG nova.scheduler.client.report [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1259.438761] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.104s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1259.439322] env[61857]: DEBUG nova.compute.manager [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1259.944552] env[61857]: DEBUG nova.compute.utils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1259.946012] env[61857]: DEBUG nova.compute.manager [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1259.946691] env[61857]: DEBUG nova.network.neutron [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1259.994755] env[61857]: DEBUG nova.policy [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4c9ee17921646978e8b21ad5508813c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2a7d2f228f88485f863b3c421b25a37c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1260.269450] env[61857]: DEBUG nova.network.neutron [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Successfully created port: 1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1260.450093] env[61857]: DEBUG nova.compute.manager [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1260.689839] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.460050] env[61857]: DEBUG nova.compute.manager [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1261.484684] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1261.484937] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1261.485125] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1261.485340] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1261.485496] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1261.485653] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1261.485864] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1261.486045] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1261.486223] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1261.486391] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1261.486571] env[61857]: DEBUG nova.virt.hardware [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1261.487456] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ea7604-321c-4fae-a278-2e6505d2b43a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.495365] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db456095-38c0-4e73-9d21-bf17dcbf3a1a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.817175] env[61857]: DEBUG nova.compute.manager [req-3b00ea7c-9b63-4e88-afd9-6694bf137316 req-857afb1f-7a98-4d59-a921-7d2903bfa093 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received event network-vif-plugged-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1261.817378] env[61857]: DEBUG oslo_concurrency.lockutils [req-3b00ea7c-9b63-4e88-afd9-6694bf137316 req-857afb1f-7a98-4d59-a921-7d2903bfa093 service nova] Acquiring lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.817610] env[61857]: DEBUG oslo_concurrency.lockutils [req-3b00ea7c-9b63-4e88-afd9-6694bf137316 req-857afb1f-7a98-4d59-a921-7d2903bfa093 service nova] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1261.817788] env[61857]: DEBUG oslo_concurrency.lockutils [req-3b00ea7c-9b63-4e88-afd9-6694bf137316 req-857afb1f-7a98-4d59-a921-7d2903bfa093 service nova] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.817967] env[61857]: DEBUG nova.compute.manager [req-3b00ea7c-9b63-4e88-afd9-6694bf137316 req-857afb1f-7a98-4d59-a921-7d2903bfa093 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] No waiting events found dispatching network-vif-plugged-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1261.818648] env[61857]: WARNING nova.compute.manager [req-3b00ea7c-9b63-4e88-afd9-6694bf137316 req-857afb1f-7a98-4d59-a921-7d2903bfa093 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received unexpected event network-vif-plugged-1978e2fa-9597-49cc-b91f-e0f550ea4007 for instance with vm_state building and task_state spawning. [ 1261.895150] env[61857]: DEBUG nova.network.neutron [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Successfully updated port: 1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1262.398452] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1262.398452] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1262.398452] env[61857]: DEBUG nova.network.neutron [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1262.929306] env[61857]: DEBUG nova.network.neutron [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1263.048470] env[61857]: DEBUG nova.network.neutron [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating instance_info_cache with network_info: [{"id": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "address": "fa:16:3e:2d:a0:f5", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1978e2fa-95", "ovs_interfaceid": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1263.551155] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1263.551460] env[61857]: DEBUG nova.compute.manager [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Instance network_info: |[{"id": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "address": "fa:16:3e:2d:a0:f5", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1978e2fa-95", "ovs_interfaceid": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1263.551940] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:a0:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '170f3b82-5915-4e36-bce9-4664ebb6be5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1978e2fa-9597-49cc-b91f-e0f550ea4007', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1263.559246] env[61857]: DEBUG oslo.service.loopingcall [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1263.559459] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1263.559718] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-863e0aff-1b35-4c28-ae70-9a0fe10a0871 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.580591] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1263.580591] env[61857]: value = "task-951740" [ 1263.580591] env[61857]: _type = "Task" [ 1263.580591] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.589418] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951740, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.685230] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1263.688960] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1263.841298] env[61857]: DEBUG nova.compute.manager [req-e472e7ba-53a6-4a57-b95c-0bd3ade674fb req-a1d6bbb9-8f46-4bbb-baa0-eb592cdcd953 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received event network-changed-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1263.841509] env[61857]: DEBUG nova.compute.manager [req-e472e7ba-53a6-4a57-b95c-0bd3ade674fb req-a1d6bbb9-8f46-4bbb-baa0-eb592cdcd953 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Refreshing instance network info cache due to event network-changed-1978e2fa-9597-49cc-b91f-e0f550ea4007. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1263.841728] env[61857]: DEBUG oslo_concurrency.lockutils [req-e472e7ba-53a6-4a57-b95c-0bd3ade674fb req-a1d6bbb9-8f46-4bbb-baa0-eb592cdcd953 service nova] Acquiring lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1263.841877] env[61857]: DEBUG oslo_concurrency.lockutils [req-e472e7ba-53a6-4a57-b95c-0bd3ade674fb req-a1d6bbb9-8f46-4bbb-baa0-eb592cdcd953 service nova] Acquired lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1263.842060] env[61857]: DEBUG nova.network.neutron [req-e472e7ba-53a6-4a57-b95c-0bd3ade674fb req-a1d6bbb9-8f46-4bbb-baa0-eb592cdcd953 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Refreshing network info cache for port 1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1264.090335] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951740, 'name': CreateVM_Task, 'duration_secs': 0.283692} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.090674] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1264.091245] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1264.091451] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.091785] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1264.092051] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b83bf65-57a4-4c8f-a367-756ac031edb6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.096598] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1264.096598] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52182de4-4a8b-9fa9-3a22-1bc4d4589489" [ 1264.096598] env[61857]: _type = "Task" [ 1264.096598] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.160587] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52182de4-4a8b-9fa9-3a22-1bc4d4589489, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.559676] env[61857]: DEBUG nova.network.neutron [req-e472e7ba-53a6-4a57-b95c-0bd3ade674fb req-a1d6bbb9-8f46-4bbb-baa0-eb592cdcd953 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updated VIF entry in instance network info cache for port 1978e2fa-9597-49cc-b91f-e0f550ea4007. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1264.560032] env[61857]: DEBUG nova.network.neutron [req-e472e7ba-53a6-4a57-b95c-0bd3ade674fb req-a1d6bbb9-8f46-4bbb-baa0-eb592cdcd953 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating instance_info_cache with network_info: [{"id": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "address": "fa:16:3e:2d:a0:f5", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1978e2fa-95", "ovs_interfaceid": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.606821] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52182de4-4a8b-9fa9-3a22-1bc4d4589489, 'name': SearchDatastore_Task, 'duration_secs': 0.009497} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.607099] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1264.607339] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1264.607577] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1264.607734] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.607919] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1264.608200] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c7ecff8f-02a8-4a95-9fb3-bf579cb8d421 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.615911] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1264.616108] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1264.616775] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1de3077-1ec8-4fef-bbfd-6c06bf4d24fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.621584] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1264.621584] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]528d83ef-9e16-ebd4-df6d-ff051cfed8bf" [ 1264.621584] env[61857]: _type = "Task" [ 1264.621584] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.628259] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528d83ef-9e16-ebd4-df6d-ff051cfed8bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.062941] env[61857]: DEBUG oslo_concurrency.lockutils [req-e472e7ba-53a6-4a57-b95c-0bd3ade674fb req-a1d6bbb9-8f46-4bbb-baa0-eb592cdcd953 service nova] Releasing lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1265.132380] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]528d83ef-9e16-ebd4-df6d-ff051cfed8bf, 'name': SearchDatastore_Task, 'duration_secs': 0.007076} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.133147] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-481a8a04-9e8e-458a-b04d-e32d7fdb891c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.138296] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1265.138296] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]5284bcf4-1d03-4673-552a-894056daa3f1" [ 1265.138296] env[61857]: _type = "Task" [ 1265.138296] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.145328] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5284bcf4-1d03-4673-552a-894056daa3f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.648190] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]5284bcf4-1d03-4673-552a-894056daa3f1, 'name': SearchDatastore_Task, 'duration_secs': 0.009113} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.648473] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1265.648735] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] c8acbb20-a33f-47c3-a37f-d350e1e21880/c8acbb20-a33f-47c3-a37f-d350e1e21880.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1265.648988] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-61705ada-7df8-482c-87ca-9b8fa3c18556 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.655902] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1265.655902] env[61857]: value = "task-951741" [ 1265.655902] env[61857]: _type = "Task" [ 1265.655902] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.664198] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951741, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.688792] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.688947] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1265.689090] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Rebuilding the list of instances to heal {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1266.166645] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951741, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.39329} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.167035] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] c8acbb20-a33f-47c3-a37f-d350e1e21880/c8acbb20-a33f-47c3-a37f-d350e1e21880.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1266.167170] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1266.167397] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca85ea85-3df1-4971-a58a-4e820abf9ec1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.173152] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1266.173152] env[61857]: value = "task-951742" [ 1266.173152] env[61857]: _type = "Task" [ 1266.173152] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.180616] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951742, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.192486] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Skipping network cache update for instance because it is Building. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1266.192623] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Didn't find any instances for network info cache update. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1266.192844] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.683567] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951742, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058907} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.683851] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1266.684695] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f5765c-7053-4cf5-8268-d5ac3219d861 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.697985] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.706652] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] c8acbb20-a33f-47c3-a37f-d350e1e21880/c8acbb20-a33f-47c3-a37f-d350e1e21880.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1266.706971] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf876f54-fe84-4bce-9fcb-0f44b11de16e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.725504] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1266.725504] env[61857]: value = "task-951743" [ 1266.725504] env[61857]: _type = "Task" [ 1266.725504] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.733412] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951743, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.235394] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951743, 'name': ReconfigVM_Task, 'duration_secs': 0.255881} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.235752] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Reconfigured VM instance instance-0000006f to attach disk [datastore2] c8acbb20-a33f-47c3-a37f-d350e1e21880/c8acbb20-a33f-47c3-a37f-d350e1e21880.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1267.236291] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c64194e6-fec4-4bdf-9e0c-257ac0b229dc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.242058] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1267.242058] env[61857]: value = "task-951744" [ 1267.242058] env[61857]: _type = "Task" [ 1267.242058] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.249448] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951744, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.698590] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1267.752047] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951744, 'name': Rename_Task, 'duration_secs': 0.133018} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.752464] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1267.752937] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e65b2117-d228-4d8c-bc6c-965750cd70ba {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.758990] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1267.758990] env[61857]: value = "task-951745" [ 1267.758990] env[61857]: _type = "Task" [ 1267.758990] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.766237] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951745, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.268579] env[61857]: DEBUG oslo_vmware.api [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951745, 'name': PowerOnVM_Task, 'duration_secs': 0.402138} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.268948] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1268.269087] env[61857]: INFO nova.compute.manager [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Took 6.81 seconds to spawn the instance on the hypervisor. [ 1268.269286] env[61857]: DEBUG nova.compute.manager [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1268.270060] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36c3bbb-8c35-4bee-ab9d-b0dcfffccc7b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.689675] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.787465] env[61857]: INFO nova.compute.manager [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Took 11.47 seconds to build instance. [ 1269.192647] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.192913] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.193100] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.193259] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1269.195821] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0f5444-088a-4b63-83ef-b1ac90177f6b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.203863] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4f465f-5256-412b-aed5-1b03e3573ea3 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.217144] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae9123e-7b33-40a8-9e22-12a26fe0b6fc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.223216] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85261dc8-7ffe-4c0d-bff3-2bf0079dd480 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.250807] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181418MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1269.250944] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.251134] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.289111] env[61857]: DEBUG oslo_concurrency.lockutils [None req-d50235d4-2ba0-4ecf-b82d-f76e85b578d2 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.979s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.388251] env[61857]: DEBUG nova.compute.manager [req-15f2b4c3-1c47-4644-95fa-cd06322b0ed8 req-f633d136-283c-4729-8d8a-954432e53dea service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received event network-changed-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1269.388490] env[61857]: DEBUG nova.compute.manager [req-15f2b4c3-1c47-4644-95fa-cd06322b0ed8 req-f633d136-283c-4729-8d8a-954432e53dea service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Refreshing instance network info cache due to event network-changed-1978e2fa-9597-49cc-b91f-e0f550ea4007. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1269.388717] env[61857]: DEBUG oslo_concurrency.lockutils [req-15f2b4c3-1c47-4644-95fa-cd06322b0ed8 req-f633d136-283c-4729-8d8a-954432e53dea service nova] Acquiring lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1269.388871] env[61857]: DEBUG oslo_concurrency.lockutils [req-15f2b4c3-1c47-4644-95fa-cd06322b0ed8 req-f633d136-283c-4729-8d8a-954432e53dea service nova] Acquired lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1269.389095] env[61857]: DEBUG nova.network.neutron [req-15f2b4c3-1c47-4644-95fa-cd06322b0ed8 req-f633d136-283c-4729-8d8a-954432e53dea service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Refreshing network info cache for port 1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1270.091851] env[61857]: DEBUG nova.network.neutron [req-15f2b4c3-1c47-4644-95fa-cd06322b0ed8 req-f633d136-283c-4729-8d8a-954432e53dea service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updated VIF entry in instance network info cache for port 1978e2fa-9597-49cc-b91f-e0f550ea4007. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1270.092236] env[61857]: DEBUG nova.network.neutron [req-15f2b4c3-1c47-4644-95fa-cd06322b0ed8 req-f633d136-283c-4729-8d8a-954432e53dea service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating instance_info_cache with network_info: [{"id": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "address": "fa:16:3e:2d:a0:f5", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1978e2fa-95", "ovs_interfaceid": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1270.273907] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c8acbb20-a33f-47c3-a37f-d350e1e21880 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1270.274166] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1270.274319] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1270.301708] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-848576de-a7c0-4779-b541-9d0709f3efd5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.309122] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51000ee4-d9a1-41cc-a320-ee03a751d939 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.339762] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b856d9-68ee-4285-a09b-389fb331ebd0 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.347190] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec0799b-0514-4a18-ad7e-1b54ca0c1e73 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.360082] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1270.595578] env[61857]: DEBUG oslo_concurrency.lockutils [req-15f2b4c3-1c47-4644-95fa-cd06322b0ed8 req-f633d136-283c-4729-8d8a-954432e53dea service nova] Releasing lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1270.863791] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1271.368581] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1271.368891] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.118s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.368932] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1273.369249] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1304.662958] env[61857]: DEBUG oslo_concurrency.lockutils [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1304.663369] env[61857]: DEBUG oslo_concurrency.lockutils [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1304.663369] env[61857]: INFO nova.compute.manager [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Shelving [ 1305.173751] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1305.174018] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-871e3d7b-7883-48e0-b9fe-3159071e1a21 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.182501] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1305.182501] env[61857]: value = "task-951746" [ 1305.182501] env[61857]: _type = "Task" [ 1305.182501] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.191059] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.692722] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951746, 'name': PowerOffVM_Task, 'duration_secs': 0.160593} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.693090] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1305.693699] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66674f94-c7d3-4221-b7d9-9d4a3dcfa7b5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.711195] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992940f7-efc2-4ac3-b489-c242ada0eaca {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.220611] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Creating Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1306.220937] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-301886b5-02cb-4769-a17e-723f99e786e2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.227911] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1306.227911] env[61857]: value = "task-951747" [ 1306.227911] env[61857]: _type = "Task" [ 1306.227911] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.236038] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951747, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.738223] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951747, 'name': CreateSnapshot_Task, 'duration_secs': 0.452037} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.738642] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Created Snapshot of the VM instance {{(pid=61857) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1306.739275] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a38007-94fd-45dd-8906-c38893c5c291 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.256717] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Creating linked-clone VM from snapshot {{(pid=61857) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1307.257047] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8f51ef3e-db67-40fe-a4a1-c9104e7ff6ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.265774] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1307.265774] env[61857]: value = "task-951748" [ 1307.265774] env[61857]: _type = "Task" [ 1307.265774] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.273262] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951748, 'name': CloneVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.775897] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951748, 'name': CloneVM_Task} progress is 94%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.276522] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951748, 'name': CloneVM_Task, 'duration_secs': 0.879251} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.276789] env[61857]: INFO nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Created linked-clone VM from snapshot [ 1308.277507] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac1c1da-8c11-4eda-abb7-3ba26e6a8908 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.285570] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Uploading image 822facc3-bd81-457f-b29b-a05a1c72bc65 {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1308.305570] env[61857]: DEBUG oslo_vmware.rw_handles [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1308.305570] env[61857]: value = "vm-214240" [ 1308.305570] env[61857]: _type = "VirtualMachine" [ 1308.305570] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1308.305793] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-95da2f40-7e67-4cd8-ae4a-ad7983890aa5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.311945] env[61857]: DEBUG oslo_vmware.rw_handles [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lease: (returnval){ [ 1308.311945] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521483ff-0113-4499-4b72-ede91cc93f75" [ 1308.311945] env[61857]: _type = "HttpNfcLease" [ 1308.311945] env[61857]: } obtained for exporting VM: (result){ [ 1308.311945] env[61857]: value = "vm-214240" [ 1308.311945] env[61857]: _type = "VirtualMachine" [ 1308.311945] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1308.312236] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the lease: (returnval){ [ 1308.312236] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521483ff-0113-4499-4b72-ede91cc93f75" [ 1308.312236] env[61857]: _type = "HttpNfcLease" [ 1308.312236] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1308.317463] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1308.317463] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521483ff-0113-4499-4b72-ede91cc93f75" [ 1308.317463] env[61857]: _type = "HttpNfcLease" [ 1308.317463] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1308.820601] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1308.820601] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521483ff-0113-4499-4b72-ede91cc93f75" [ 1308.820601] env[61857]: _type = "HttpNfcLease" [ 1308.820601] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1308.821055] env[61857]: DEBUG oslo_vmware.rw_handles [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1308.821055] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]521483ff-0113-4499-4b72-ede91cc93f75" [ 1308.821055] env[61857]: _type = "HttpNfcLease" [ 1308.821055] env[61857]: }. {{(pid=61857) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1308.821620] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e943e2-c60c-4cf6-9910-e286ba4d992c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.828607] env[61857]: DEBUG oslo_vmware.rw_handles [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e91371-826f-19f8-bf6a-bf106b9fa40b/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1308.828784] env[61857]: DEBUG oslo_vmware.rw_handles [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e91371-826f-19f8-bf6a-bf106b9fa40b/disk-0.vmdk for reading. {{(pid=61857) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1308.914104] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-af0ebdd9-b916-4a18-a35d-7f9f244c8917 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.702989] env[61857]: DEBUG oslo_vmware.rw_handles [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e91371-826f-19f8-bf6a-bf106b9fa40b/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1316.703911] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27c6d7d-7191-408f-bb2d-20e554a38c94 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.709903] env[61857]: DEBUG oslo_vmware.rw_handles [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e91371-826f-19f8-bf6a-bf106b9fa40b/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1316.710090] env[61857]: ERROR oslo_vmware.rw_handles [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e91371-826f-19f8-bf6a-bf106b9fa40b/disk-0.vmdk due to incomplete transfer. [ 1316.710301] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-197f66bd-dda5-4632-99dc-3d676897a4c4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.716269] env[61857]: DEBUG oslo_vmware.rw_handles [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e91371-826f-19f8-bf6a-bf106b9fa40b/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1316.716466] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Uploaded image 822facc3-bd81-457f-b29b-a05a1c72bc65 to the Glance image server {{(pid=61857) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1316.718802] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Destroying the VM {{(pid=61857) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1316.719036] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-767d77c8-9f10-4882-a0c1-ade42e4717cb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.724015] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1316.724015] env[61857]: value = "task-951750" [ 1316.724015] env[61857]: _type = "Task" [ 1316.724015] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.731152] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951750, 'name': Destroy_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.233775] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951750, 'name': Destroy_Task, 'duration_secs': 0.294504} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.234050] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Destroyed the VM [ 1317.234299] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Deleting Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1317.234560] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f1d5d514-1e49-4329-8fbe-dae18b70eb13 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.240919] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1317.240919] env[61857]: value = "task-951751" [ 1317.240919] env[61857]: _type = "Task" [ 1317.240919] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1317.247980] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951751, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.751405] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951751, 'name': RemoveSnapshot_Task, 'duration_secs': 0.315855} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.751737] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Deleted Snapshot of the VM instance {{(pid=61857) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1317.751950] env[61857]: DEBUG nova.compute.manager [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1317.752703] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09487607-1391-407a-ada2-12b415154ad7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.264200] env[61857]: INFO nova.compute.manager [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Shelve offloading [ 1318.265805] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1318.266056] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c264dbb-9909-4aab-91d8-09b78ef99767 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.273472] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1318.273472] env[61857]: value = "task-951752" [ 1318.273472] env[61857]: _type = "Task" [ 1318.273472] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.280951] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951752, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.785419] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] VM already powered off {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1318.785786] env[61857]: DEBUG nova.compute.manager [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1318.786480] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5638af77-919b-45fc-aaba-86c10e1bfb7e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.792400] env[61857]: DEBUG oslo_concurrency.lockutils [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1318.792571] env[61857]: DEBUG oslo_concurrency.lockutils [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1318.792742] env[61857]: DEBUG nova.network.neutron [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1319.496916] env[61857]: DEBUG nova.network.neutron [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating instance_info_cache with network_info: [{"id": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "address": "fa:16:3e:2d:a0:f5", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1978e2fa-95", "ovs_interfaceid": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1319.999906] env[61857]: DEBUG oslo_concurrency.lockutils [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1320.208787] env[61857]: DEBUG nova.compute.manager [req-bd909775-c593-42cf-9eea-b6367a41fdf0 req-988004da-c0db-4742-8089-28cea79e00f0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received event network-vif-unplugged-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1320.209035] env[61857]: DEBUG oslo_concurrency.lockutils [req-bd909775-c593-42cf-9eea-b6367a41fdf0 req-988004da-c0db-4742-8089-28cea79e00f0 service nova] Acquiring lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1320.209310] env[61857]: DEBUG oslo_concurrency.lockutils [req-bd909775-c593-42cf-9eea-b6367a41fdf0 req-988004da-c0db-4742-8089-28cea79e00f0 service nova] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1320.209438] env[61857]: DEBUG oslo_concurrency.lockutils [req-bd909775-c593-42cf-9eea-b6367a41fdf0 req-988004da-c0db-4742-8089-28cea79e00f0 service nova] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1320.209572] env[61857]: DEBUG nova.compute.manager [req-bd909775-c593-42cf-9eea-b6367a41fdf0 req-988004da-c0db-4742-8089-28cea79e00f0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] No waiting events found dispatching network-vif-unplugged-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1320.209752] env[61857]: WARNING nova.compute.manager [req-bd909775-c593-42cf-9eea-b6367a41fdf0 req-988004da-c0db-4742-8089-28cea79e00f0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received unexpected event network-vif-unplugged-1978e2fa-9597-49cc-b91f-e0f550ea4007 for instance with vm_state shelved and task_state shelving_offloading. [ 1320.301644] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1320.302558] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c919ea90-ecea-4365-9007-558015ec3297 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.310015] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1320.310251] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-110e47dd-b8ab-4193-9138-8c1add5800f4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.380451] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1320.380716] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1320.380918] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleting the datastore file [datastore2] c8acbb20-a33f-47c3-a37f-d350e1e21880 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1320.381207] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6756da4c-3145-4c4a-bb6f-bdf292f15daf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.387530] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1320.387530] env[61857]: value = "task-951754" [ 1320.387530] env[61857]: _type = "Task" [ 1320.387530] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.394926] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951754, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.689975] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.896861] env[61857]: DEBUG oslo_vmware.api [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951754, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121988} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.897154] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1320.897327] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1320.897505] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1320.920173] env[61857]: INFO nova.scheduler.client.report [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleted allocations for instance c8acbb20-a33f-47c3-a37f-d350e1e21880 [ 1321.424915] env[61857]: DEBUG oslo_concurrency.lockutils [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1321.425257] env[61857]: DEBUG oslo_concurrency.lockutils [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1321.425512] env[61857]: DEBUG nova.objects.instance [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'resources' on Instance uuid c8acbb20-a33f-47c3-a37f-d350e1e21880 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1321.929394] env[61857]: DEBUG nova.objects.instance [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'numa_topology' on Instance uuid c8acbb20-a33f-47c3-a37f-d350e1e21880 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1322.239183] env[61857]: DEBUG nova.compute.manager [req-473775e1-8416-47ac-883b-d71ed2cc7a97 req-b4132599-8d3f-4e08-8cbd-7322fc00dcf0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received event network-changed-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1322.239399] env[61857]: DEBUG nova.compute.manager [req-473775e1-8416-47ac-883b-d71ed2cc7a97 req-b4132599-8d3f-4e08-8cbd-7322fc00dcf0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Refreshing instance network info cache due to event network-changed-1978e2fa-9597-49cc-b91f-e0f550ea4007. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1322.239620] env[61857]: DEBUG oslo_concurrency.lockutils [req-473775e1-8416-47ac-883b-d71ed2cc7a97 req-b4132599-8d3f-4e08-8cbd-7322fc00dcf0 service nova] Acquiring lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1322.239770] env[61857]: DEBUG oslo_concurrency.lockutils [req-473775e1-8416-47ac-883b-d71ed2cc7a97 req-b4132599-8d3f-4e08-8cbd-7322fc00dcf0 service nova] Acquired lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1322.239937] env[61857]: DEBUG nova.network.neutron [req-473775e1-8416-47ac-883b-d71ed2cc7a97 req-b4132599-8d3f-4e08-8cbd-7322fc00dcf0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Refreshing network info cache for port 1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1322.431586] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1322.431989] env[61857]: DEBUG nova.objects.base [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61857) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1322.446620] env[61857]: DEBUG nova.scheduler.client.report [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Refreshing inventories for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1322.459304] env[61857]: DEBUG nova.scheduler.client.report [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Updating ProviderTree inventory for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1322.459521] env[61857]: DEBUG nova.compute.provider_tree [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Updating inventory in ProviderTree for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1322.469504] env[61857]: DEBUG nova.scheduler.client.report [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Refreshing aggregate associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, aggregates: None {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1322.483606] env[61857]: DEBUG nova.scheduler.client.report [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Refreshing trait associations for resource provider 2d5860d0-8574-4e55-9ced-20e66f0314c2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61857) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1322.504578] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19caabf8-85fe-4d04-b748-8cecd4eb70af {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.512211] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d2e96d-43f1-45a6-b5e1-1716a0314281 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.542271] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a69a7d5-f582-4726-9785-7899dccad3fd {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.549188] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97cf32ea-91c8-4213-9d69-d854f57c8b2e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.561930] env[61857]: DEBUG nova.compute.provider_tree [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1322.949197] env[61857]: DEBUG nova.network.neutron [req-473775e1-8416-47ac-883b-d71ed2cc7a97 req-b4132599-8d3f-4e08-8cbd-7322fc00dcf0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updated VIF entry in instance network info cache for port 1978e2fa-9597-49cc-b91f-e0f550ea4007. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1322.949581] env[61857]: DEBUG nova.network.neutron [req-473775e1-8416-47ac-883b-d71ed2cc7a97 req-b4132599-8d3f-4e08-8cbd-7322fc00dcf0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating instance_info_cache with network_info: [{"id": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "address": "fa:16:3e:2d:a0:f5", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap1978e2fa-95", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1323.064513] env[61857]: DEBUG nova.scheduler.client.report [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1323.452588] env[61857]: DEBUG oslo_concurrency.lockutils [req-473775e1-8416-47ac-883b-d71ed2cc7a97 req-b4132599-8d3f-4e08-8cbd-7322fc00dcf0 service nova] Releasing lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1323.568948] env[61857]: DEBUG oslo_concurrency.lockutils [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.144s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1323.689964] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1324.076702] env[61857]: DEBUG oslo_concurrency.lockutils [None req-749e657e-8174-4d2a-870b-8ed78a8bee83 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.413s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1324.077513] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.646s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1324.077701] env[61857]: INFO nova.compute.manager [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Unshelving [ 1324.685070] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.099110] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.099402] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.099629] env[61857]: DEBUG nova.objects.instance [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'pci_requests' on Instance uuid c8acbb20-a33f-47c3-a37f-d350e1e21880 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1325.604026] env[61857]: DEBUG nova.objects.instance [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'numa_topology' on Instance uuid c8acbb20-a33f-47c3-a37f-d350e1e21880 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1326.107179] env[61857]: INFO nova.compute.claims [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1327.141892] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2847579c-721a-48d2-976f-7e2cec206cd4 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.149243] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea7585e-7615-4190-bfa9-bd20d623a247 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.178376] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8079d5f8-dd0d-4533-9a8f-ac3ab1c6aa5b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.185254] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b17cbd6-bb30-43ea-b391-ff6fd2d28de1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.197877] env[61857]: DEBUG nova.compute.provider_tree [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1327.688990] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1327.689211] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1327.689335] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Rebuilding the list of instances to heal {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1327.700661] env[61857]: DEBUG nova.scheduler.client.report [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1328.192294] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1328.192652] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquired lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.192652] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Forcefully refreshing network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1328.192792] env[61857]: DEBUG nova.objects.instance [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lazy-loading 'info_cache' on Instance uuid c8acbb20-a33f-47c3-a37f-d350e1e21880 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1328.205046] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.106s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1328.232210] env[61857]: INFO nova.network.neutron [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating port 1978e2fa-9597-49cc-b91f-e0f550ea4007 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1329.901235] env[61857]: DEBUG nova.network.neutron [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating instance_info_cache with network_info: [{"id": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "address": "fa:16:3e:2d:a0:f5", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap1978e2fa-95", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1330.097737] env[61857]: DEBUG nova.compute.manager [req-7a1e7ac7-fd8f-49cc-bdfc-151b5c8d68df req-8060dd94-9648-4edc-90a0-f3c83be7e5e0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received event network-vif-plugged-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1330.097949] env[61857]: DEBUG oslo_concurrency.lockutils [req-7a1e7ac7-fd8f-49cc-bdfc-151b5c8d68df req-8060dd94-9648-4edc-90a0-f3c83be7e5e0 service nova] Acquiring lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.098183] env[61857]: DEBUG oslo_concurrency.lockutils [req-7a1e7ac7-fd8f-49cc-bdfc-151b5c8d68df req-8060dd94-9648-4edc-90a0-f3c83be7e5e0 service nova] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.098356] env[61857]: DEBUG oslo_concurrency.lockutils [req-7a1e7ac7-fd8f-49cc-bdfc-151b5c8d68df req-8060dd94-9648-4edc-90a0-f3c83be7e5e0 service nova] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.098558] env[61857]: DEBUG nova.compute.manager [req-7a1e7ac7-fd8f-49cc-bdfc-151b5c8d68df req-8060dd94-9648-4edc-90a0-f3c83be7e5e0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] No waiting events found dispatching network-vif-plugged-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1330.098733] env[61857]: WARNING nova.compute.manager [req-7a1e7ac7-fd8f-49cc-bdfc-151b5c8d68df req-8060dd94-9648-4edc-90a0-f3c83be7e5e0 service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received unexpected event network-vif-plugged-1978e2fa-9597-49cc-b91f-e0f550ea4007 for instance with vm_state shelved_offloaded and task_state spawning. [ 1330.185765] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1330.403949] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Releasing lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1330.404244] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updated the network info_cache for instance {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1330.404563] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1330.404789] env[61857]: DEBUG nova.network.neutron [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1330.406007] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.406207] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.406371] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.689016] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1331.098570] env[61857]: DEBUG nova.network.neutron [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating instance_info_cache with network_info: [{"id": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "address": "fa:16:3e:2d:a0:f5", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1978e2fa-95", "ovs_interfaceid": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1331.191498] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.191735] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.191901] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1331.192065] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1331.193013] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059eb6f6-8483-4d89-954c-c61dee3ff3d5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.201196] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d73218-e6e9-49f5-a946-34c0d61ef418 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.216059] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228014dc-d3a7-4a32-bbfa-89fb79a2a5c1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.221977] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b75aee-a035-4d90-8baa-51f8232ad113 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.251534] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181628MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1331.251671] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.251890] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.601926] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1331.630931] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='164090fb3b8bd550655d318be65a5e80',container_format='bare',created_at=2024-09-18T21:51:07Z,direct_url=,disk_format='vmdk',id=822facc3-bd81-457f-b29b-a05a1c72bc65,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1868590964-shelved',owner='2a7d2f228f88485f863b3c421b25a37c',properties=ImageMetaProps,protected=,size=31670784,status='active',tags=,updated_at=2024-09-18T21:51:20Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1331.631199] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1331.631373] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1331.631626] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1331.631807] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1331.631966] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1331.632197] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1331.632367] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1331.632541] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1331.632714] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1331.632893] env[61857]: DEBUG nova.virt.hardware [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1331.633744] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ceba16-9017-49dd-bf35-c7cf93718682 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.641266] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dc92d0-39c8-46a9-87b2-25ec5bf813df {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.654431] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:a0:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '170f3b82-5915-4e36-bce9-4664ebb6be5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1978e2fa-9597-49cc-b91f-e0f550ea4007', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1331.661824] env[61857]: DEBUG oslo.service.loopingcall [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1331.662071] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1331.662285] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4b96da0-2075-48c7-8087-4a74b0bef1bf {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.680975] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1331.680975] env[61857]: value = "task-951755" [ 1331.680975] env[61857]: _type = "Task" [ 1331.680975] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.688259] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951755, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.122878] env[61857]: DEBUG nova.compute.manager [req-fb0dd111-9d1c-4450-993d-26f47ef7c76f req-b340b39a-83fc-4418-bab4-50fe40af725f service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received event network-changed-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1332.123140] env[61857]: DEBUG nova.compute.manager [req-fb0dd111-9d1c-4450-993d-26f47ef7c76f req-b340b39a-83fc-4418-bab4-50fe40af725f service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Refreshing instance network info cache due to event network-changed-1978e2fa-9597-49cc-b91f-e0f550ea4007. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1332.123338] env[61857]: DEBUG oslo_concurrency.lockutils [req-fb0dd111-9d1c-4450-993d-26f47ef7c76f req-b340b39a-83fc-4418-bab4-50fe40af725f service nova] Acquiring lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1332.123490] env[61857]: DEBUG oslo_concurrency.lockutils [req-fb0dd111-9d1c-4450-993d-26f47ef7c76f req-b340b39a-83fc-4418-bab4-50fe40af725f service nova] Acquired lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1332.123658] env[61857]: DEBUG nova.network.neutron [req-fb0dd111-9d1c-4450-993d-26f47ef7c76f req-b340b39a-83fc-4418-bab4-50fe40af725f service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Refreshing network info cache for port 1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1332.191313] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951755, 'name': CreateVM_Task, 'duration_secs': 0.294391} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.191504] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1332.198882] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1332.199075] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "[datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1332.199469] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1332.199739] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a309ddf9-4e8e-4265-ba82-9aaa7a86606b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.204198] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1332.204198] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52626db4-9d02-75ac-de61-983635b84be5" [ 1332.204198] env[61857]: _type = "Task" [ 1332.204198] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.211502] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52626db4-9d02-75ac-de61-983635b84be5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.271099] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Instance c8acbb20-a33f-47c3-a37f-d350e1e21880 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61857) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1332.271294] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1332.271460] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1332.294637] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e9a01b-3628-46b1-861a-9fb9478f316e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.301986] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf6e6c2-668a-4331-a5e7-fb70b68f006d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.332216] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77910aa9-7381-40bc-aedd-ca2568618a6e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.338854] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3089e6b-54b4-4caa-835e-119c9cf4c51c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.351437] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1332.716042] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "[datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1332.716323] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Processing image 822facc3-bd81-457f-b29b-a05a1c72bc65 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1332.716560] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65/822facc3-bd81-457f-b29b-a05a1c72bc65.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1332.716714] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquired lock "[datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65/822facc3-bd81-457f-b29b-a05a1c72bc65.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1332.716891] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1332.717165] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df27ac30-6303-4531-9c68-413be6e70e92 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.725597] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1332.725779] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1332.726461] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01817703-9ce7-4ed8-9fd7-aa0626896d8c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.731151] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1332.731151] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]522ddb74-1a0e-6164-600d-b1a933ef4619" [ 1332.731151] env[61857]: _type = "Task" [ 1332.731151] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.740582] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]522ddb74-1a0e-6164-600d-b1a933ef4619, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.823206] env[61857]: DEBUG nova.network.neutron [req-fb0dd111-9d1c-4450-993d-26f47ef7c76f req-b340b39a-83fc-4418-bab4-50fe40af725f service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updated VIF entry in instance network info cache for port 1978e2fa-9597-49cc-b91f-e0f550ea4007. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1332.823564] env[61857]: DEBUG nova.network.neutron [req-fb0dd111-9d1c-4450-993d-26f47ef7c76f req-b340b39a-83fc-4418-bab4-50fe40af725f service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating instance_info_cache with network_info: [{"id": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "address": "fa:16:3e:2d:a0:f5", "network": {"id": "53f035b0-5e2b-4472-9ad7-f5c084c709c1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1390310227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a7d2f228f88485f863b3c421b25a37c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1978e2fa-95", "ovs_interfaceid": "1978e2fa-9597-49cc-b91f-e0f550ea4007", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1332.854675] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1333.241386] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Preparing fetch location {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1333.241739] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Fetch image to [datastore2] OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2/OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2.vmdk {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1333.241830] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Downloading stream optimized image 822facc3-bd81-457f-b29b-a05a1c72bc65 to [datastore2] OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2/OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2.vmdk on the data store datastore2 as vApp {{(pid=61857) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1333.242014] env[61857]: DEBUG nova.virt.vmwareapi.images [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Downloading image file data 822facc3-bd81-457f-b29b-a05a1c72bc65 to the ESX as VM named 'OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2' {{(pid=61857) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1333.304831] env[61857]: DEBUG oslo_vmware.rw_handles [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1333.304831] env[61857]: value = "resgroup-9" [ 1333.304831] env[61857]: _type = "ResourcePool" [ 1333.304831] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1333.305138] env[61857]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d657436e-8686-42a1-a82e-2a2a1c379eb8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.325535] env[61857]: DEBUG oslo_concurrency.lockutils [req-fb0dd111-9d1c-4450-993d-26f47ef7c76f req-b340b39a-83fc-4418-bab4-50fe40af725f service nova] Releasing lock "refresh_cache-c8acbb20-a33f-47c3-a37f-d350e1e21880" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1333.325950] env[61857]: DEBUG oslo_vmware.rw_handles [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lease: (returnval){ [ 1333.325950] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524dc1ed-4847-e305-f6a7-7e51a20ed535" [ 1333.325950] env[61857]: _type = "HttpNfcLease" [ 1333.325950] env[61857]: } obtained for vApp import into resource pool (val){ [ 1333.325950] env[61857]: value = "resgroup-9" [ 1333.325950] env[61857]: _type = "ResourcePool" [ 1333.325950] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1333.326187] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the lease: (returnval){ [ 1333.326187] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524dc1ed-4847-e305-f6a7-7e51a20ed535" [ 1333.326187] env[61857]: _type = "HttpNfcLease" [ 1333.326187] env[61857]: } to be ready. {{(pid=61857) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1333.332099] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1333.332099] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524dc1ed-4847-e305-f6a7-7e51a20ed535" [ 1333.332099] env[61857]: _type = "HttpNfcLease" [ 1333.332099] env[61857]: } is initializing. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1333.358899] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1333.359106] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.107s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1333.841744] env[61857]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1333.841744] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524dc1ed-4847-e305-f6a7-7e51a20ed535" [ 1333.841744] env[61857]: _type = "HttpNfcLease" [ 1333.841744] env[61857]: } is ready. {{(pid=61857) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1333.842075] env[61857]: DEBUG oslo_vmware.rw_handles [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1333.842075] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]524dc1ed-4847-e305-f6a7-7e51a20ed535" [ 1333.842075] env[61857]: _type = "HttpNfcLease" [ 1333.842075] env[61857]: }. {{(pid=61857) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1333.842707] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12068e82-544f-4aeb-9205-b2ce08b1f7b2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.851249] env[61857]: DEBUG oslo_vmware.rw_handles [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263229f-bfef-6178-95ab-da7140bd4279/disk-0.vmdk from lease info. {{(pid=61857) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1333.851392] env[61857]: DEBUG oslo_vmware.rw_handles [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating HTTP connection to write to file with size = 31670784 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263229f-bfef-6178-95ab-da7140bd4279/disk-0.vmdk. {{(pid=61857) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1333.914024] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1f998ebd-fa2c-4452-aae0-bd802ed6029e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.355209] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1334.861998] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1334.862264] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1334.990967] env[61857]: DEBUG oslo_vmware.rw_handles [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Completed reading data from the image iterator. {{(pid=61857) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1334.991261] env[61857]: DEBUG oslo_vmware.rw_handles [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263229f-bfef-6178-95ab-da7140bd4279/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1334.992220] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae79e0fc-f0d8-437a-a9e0-a92382e04a8c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.999346] env[61857]: DEBUG oslo_vmware.rw_handles [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263229f-bfef-6178-95ab-da7140bd4279/disk-0.vmdk is in state: ready. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1334.999517] env[61857]: DEBUG oslo_vmware.rw_handles [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263229f-bfef-6178-95ab-da7140bd4279/disk-0.vmdk. {{(pid=61857) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1334.999747] env[61857]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1fa731bb-5f14-41f8-917b-29479a6c47b2 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.183336] env[61857]: DEBUG oslo_vmware.rw_handles [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5263229f-bfef-6178-95ab-da7140bd4279/disk-0.vmdk. {{(pid=61857) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1335.183515] env[61857]: INFO nova.virt.vmwareapi.images [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Downloaded image file data 822facc3-bd81-457f-b29b-a05a1c72bc65 [ 1335.184395] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac9070c7-b32e-4249-a80f-c025e238ec49 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.199550] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66bad179-4fb7-425c-8e10-03b479ea79a9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.229820] env[61857]: INFO nova.virt.vmwareapi.images [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] The imported VM was unregistered [ 1335.232186] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Caching image {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1335.232423] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Creating directory with path [datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65 {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1335.232697] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-519ee934-86b0-4a3e-8b2e-b905d4e1df17 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.241856] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Created directory with path [datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65 {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1335.242054] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2/OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2.vmdk to [datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65/822facc3-bd81-457f-b29b-a05a1c72bc65.vmdk. {{(pid=61857) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1335.242282] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-26060a11-6730-4cb2-a5f9-523ffa764c0f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.248207] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1335.248207] env[61857]: value = "task-951758" [ 1335.248207] env[61857]: _type = "Task" [ 1335.248207] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.255495] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951758, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.758722] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951758, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.259082] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951758, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.760238] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951758, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.262422] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951758, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.762148] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951758, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.094115} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.762584] env[61857]: INFO nova.virt.vmwareapi.ds_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2/OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2.vmdk to [datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65/822facc3-bd81-457f-b29b-a05a1c72bc65.vmdk. [ 1337.762629] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Cleaning up location [datastore2] OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1337.762791] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_7c3a7780-1c77-42b3-bff5-a779bd03f6b2 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1337.763065] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58915ff9-bee2-41f8-8241-ce2eb282dc96 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.769321] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1337.769321] env[61857]: value = "task-951759" [ 1337.769321] env[61857]: _type = "Task" [ 1337.769321] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.776635] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951759, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.279058] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951759, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031178} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.279285] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1338.279458] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Releasing lock "[datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65/822facc3-bd81-457f-b29b-a05a1c72bc65.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1338.279706] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65/822facc3-bd81-457f-b29b-a05a1c72bc65.vmdk to [datastore2] c8acbb20-a33f-47c3-a37f-d350e1e21880/c8acbb20-a33f-47c3-a37f-d350e1e21880.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1338.279946] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3737e43b-4619-4c30-a5a4-a8681dec5a06 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.286093] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1338.286093] env[61857]: value = "task-951760" [ 1338.286093] env[61857]: _type = "Task" [ 1338.286093] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.293840] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951760, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.796126] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951760, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.299100] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951760, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.797962] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951760, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.299017] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951760, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.799325] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951760, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.076203} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.799711] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/822facc3-bd81-457f-b29b-a05a1c72bc65/822facc3-bd81-457f-b29b-a05a1c72bc65.vmdk to [datastore2] c8acbb20-a33f-47c3-a37f-d350e1e21880/c8acbb20-a33f-47c3-a37f-d350e1e21880.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1340.800377] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef97fe1-f450-4cb5-8401-c488aeea9d65 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.821600] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] c8acbb20-a33f-47c3-a37f-d350e1e21880/c8acbb20-a33f-47c3-a37f-d350e1e21880.vmdk or device None with type streamOptimized {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1340.821839] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38b030c0-e95a-4f90-81da-8e644b57a4ec {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.840601] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1340.840601] env[61857]: value = "task-951761" [ 1340.840601] env[61857]: _type = "Task" [ 1340.840601] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.848083] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951761, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.350528] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951761, 'name': ReconfigVM_Task, 'duration_secs': 0.27559} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.350823] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Reconfigured VM instance instance-0000006f to attach disk [datastore2] c8acbb20-a33f-47c3-a37f-d350e1e21880/c8acbb20-a33f-47c3-a37f-d350e1e21880.vmdk or device None with type streamOptimized {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1341.352027] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f70e2854-0b07-447a-89b0-918825958037 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.357995] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1341.357995] env[61857]: value = "task-951762" [ 1341.357995] env[61857]: _type = "Task" [ 1341.357995] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.365470] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951762, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.867464] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951762, 'name': Rename_Task, 'duration_secs': 0.147712} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.867836] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1341.867977] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edcc5ddb-1bfc-4fa2-bac3-a919d3fd1744 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.874211] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1341.874211] env[61857]: value = "task-951763" [ 1341.874211] env[61857]: _type = "Task" [ 1341.874211] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.881313] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951763, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.384799] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951763, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.885811] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951763, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.386458] env[61857]: DEBUG oslo_vmware.api [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951763, 'name': PowerOnVM_Task, 'duration_secs': 1.032381} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.386740] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1343.482853] env[61857]: DEBUG nova.compute.manager [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1343.483846] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b453ac-3a78-4998-81b8-1c51677d7c1e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.001142] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bb86fa8f-5920-4ca9-a729-327e2ea648af tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.924s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1344.366033] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1344.366334] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1344.366554] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1344.366748] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1344.366922] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1344.369017] env[61857]: INFO nova.compute.manager [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Terminating instance [ 1344.370736] env[61857]: DEBUG nova.compute.manager [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1344.370950] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1344.371795] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09b5de0-1547-4a13-a696-02a2c2e04dfe {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.379554] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1344.379849] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0ba82db-7a17-4b3a-8b9e-d48aae461bb7 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.385766] env[61857]: DEBUG oslo_vmware.api [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1344.385766] env[61857]: value = "task-951764" [ 1344.385766] env[61857]: _type = "Task" [ 1344.385766] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.393476] env[61857]: DEBUG oslo_vmware.api [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.896666] env[61857]: DEBUG oslo_vmware.api [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951764, 'name': PowerOffVM_Task, 'duration_secs': 0.198562} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.896993] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1344.897287] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1344.897544] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03d97eca-fbde-4e3e-bb07-6e5bf8f6e270 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.969111] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1344.970026] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1344.970026] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleting the datastore file [datastore2] c8acbb20-a33f-47c3-a37f-d350e1e21880 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1344.970026] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3fe7808f-934f-4359-a771-d7ad0549c01c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.976629] env[61857]: DEBUG oslo_vmware.api [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for the task: (returnval){ [ 1344.976629] env[61857]: value = "task-951766" [ 1344.976629] env[61857]: _type = "Task" [ 1344.976629] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.985237] env[61857]: DEBUG oslo_vmware.api [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951766, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.488692] env[61857]: DEBUG oslo_vmware.api [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Task: {'id': task-951766, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140495} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.489095] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1345.489095] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1345.489276] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1345.489457] env[61857]: INFO nova.compute.manager [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1345.489711] env[61857]: DEBUG oslo.service.loopingcall [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1345.489910] env[61857]: DEBUG nova.compute.manager [-] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1345.490017] env[61857]: DEBUG nova.network.neutron [-] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1345.928045] env[61857]: DEBUG nova.compute.manager [req-e0fbce9c-a9d5-4364-af9c-1d1ab85b22b1 req-b9e092aa-2f84-49ea-b854-8d131517514a service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Received event network-vif-deleted-1978e2fa-9597-49cc-b91f-e0f550ea4007 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1345.928356] env[61857]: INFO nova.compute.manager [req-e0fbce9c-a9d5-4364-af9c-1d1ab85b22b1 req-b9e092aa-2f84-49ea-b854-8d131517514a service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Neutron deleted interface 1978e2fa-9597-49cc-b91f-e0f550ea4007; detaching it from the instance and deleting it from the info cache [ 1345.928470] env[61857]: DEBUG nova.network.neutron [req-e0fbce9c-a9d5-4364-af9c-1d1ab85b22b1 req-b9e092aa-2f84-49ea-b854-8d131517514a service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1346.405036] env[61857]: DEBUG nova.network.neutron [-] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1346.431263] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1517e033-0870-4a94-948f-5d26c63a1a14 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.441413] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc054877-06d9-4fcd-ba2c-1591528c648e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.464440] env[61857]: DEBUG nova.compute.manager [req-e0fbce9c-a9d5-4364-af9c-1d1ab85b22b1 req-b9e092aa-2f84-49ea-b854-8d131517514a service nova] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Detach interface failed, port_id=1978e2fa-9597-49cc-b91f-e0f550ea4007, reason: Instance c8acbb20-a33f-47c3-a37f-d350e1e21880 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1346.907412] env[61857]: INFO nova.compute.manager [-] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Took 1.42 seconds to deallocate network for instance. [ 1347.414754] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1347.415120] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1347.415399] env[61857]: DEBUG nova.objects.instance [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lazy-loading 'resources' on Instance uuid c8acbb20-a33f-47c3-a37f-d350e1e21880 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1347.951755] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504b2b55-b9e0-4e2d-90c9-26000d81b52b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.959355] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a3fa89-f394-4caa-a822-525a04611653 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.988365] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944e4036-07f6-400b-bf7c-e6b009a138e5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.995436] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ffc75c7-9113-4ab0-bba5-a39409d77602 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.007946] env[61857]: DEBUG nova.compute.provider_tree [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1348.510729] env[61857]: DEBUG nova.scheduler.client.report [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1349.016311] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.601s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1349.036154] env[61857]: INFO nova.scheduler.client.report [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Deleted allocations for instance c8acbb20-a33f-47c3-a37f-d350e1e21880 [ 1349.543340] env[61857]: DEBUG oslo_concurrency.lockutils [None req-5d8a2bfb-8a13-49de-919c-2a639ee6d591 tempest-ServerActionsTestOtherB-1811840184 tempest-ServerActionsTestOtherB-1811840184-project-member] Lock "c8acbb20-a33f-47c3-a37f-d350e1e21880" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.177s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1357.385368] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquiring lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1357.385674] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1357.888180] env[61857]: DEBUG nova.compute.manager [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Starting instance... {{(pid=61857) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1358.410258] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1358.410639] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1358.412427] env[61857]: INFO nova.compute.claims [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1359.448620] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabf9616-2523-4fca-8d1d-c4d08e33c7f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.456488] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8fbc71-86a1-414e-9e09-89e2b7d5c933 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.486157] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0f1de9-b75c-4dc5-aafe-48e7246d384d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.493810] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316641ed-44d5-428c-8250-1e5bab87617a {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.508279] env[61857]: DEBUG nova.compute.provider_tree [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1360.011475] env[61857]: DEBUG nova.scheduler.client.report [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1360.516072] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.105s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1360.516640] env[61857]: DEBUG nova.compute.manager [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Start building networks asynchronously for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1361.021800] env[61857]: DEBUG nova.compute.utils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Using /dev/sd instead of None {{(pid=61857) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1361.023233] env[61857]: DEBUG nova.compute.manager [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Allocating IP information in the background. {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1361.023406] env[61857]: DEBUG nova.network.neutron [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] allocate_for_instance() {{(pid=61857) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1361.059801] env[61857]: DEBUG nova.policy [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb5ccd666d5743e6b2323479e6df3471', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b3f950eeb3b468ca7cf97066329f564', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61857) authorize /opt/stack/nova/nova/policy.py:201}} [ 1361.326078] env[61857]: DEBUG nova.network.neutron [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Successfully created port: f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870 {{(pid=61857) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1361.527324] env[61857]: DEBUG nova.compute.manager [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Start building block device mappings for instance. {{(pid=61857) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1362.535872] env[61857]: DEBUG nova.compute.manager [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Start spawning the instance on the hypervisor. {{(pid=61857) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1362.560213] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T21:36:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T21:36:12Z,direct_url=,disk_format='vmdk',id=2e0c8bb4-0827-486c-98ce-7994555ffc23,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='fabda895e0be4fd3b4a2ed5c7cda3a13',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T21:36:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1362.560506] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Flavor limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1362.560684] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Image limits 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1362.560876] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Flavor pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1362.561043] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Image pref 0:0:0 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1362.561202] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61857) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1362.561415] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1362.561582] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1362.561750] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Got 1 possible topologies {{(pid=61857) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1362.561914] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1362.562525] env[61857]: DEBUG nova.virt.hardware [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61857) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1362.562976] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5494f8-830a-458d-83f5-c336d5542d68 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.570747] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e055e5-1b56-4454-b0e4-dd6065095dcb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.681720] env[61857]: DEBUG nova.compute.manager [req-47e9b74e-1016-466c-a441-ef90fd4bbe75 req-f2806d59-9594-45a8-a408-2c798ccadeb7 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Received event network-vif-plugged-f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1362.681775] env[61857]: DEBUG oslo_concurrency.lockutils [req-47e9b74e-1016-466c-a441-ef90fd4bbe75 req-f2806d59-9594-45a8-a408-2c798ccadeb7 service nova] Acquiring lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1362.681971] env[61857]: DEBUG oslo_concurrency.lockutils [req-47e9b74e-1016-466c-a441-ef90fd4bbe75 req-f2806d59-9594-45a8-a408-2c798ccadeb7 service nova] Lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1362.682164] env[61857]: DEBUG oslo_concurrency.lockutils [req-47e9b74e-1016-466c-a441-ef90fd4bbe75 req-f2806d59-9594-45a8-a408-2c798ccadeb7 service nova] Lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1362.682341] env[61857]: DEBUG nova.compute.manager [req-47e9b74e-1016-466c-a441-ef90fd4bbe75 req-f2806d59-9594-45a8-a408-2c798ccadeb7 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] No waiting events found dispatching network-vif-plugged-f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870 {{(pid=61857) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1362.682516] env[61857]: WARNING nova.compute.manager [req-47e9b74e-1016-466c-a441-ef90fd4bbe75 req-f2806d59-9594-45a8-a408-2c798ccadeb7 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Received unexpected event network-vif-plugged-f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870 for instance with vm_state building and task_state spawning. [ 1362.764244] env[61857]: DEBUG nova.network.neutron [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Successfully updated port: f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870 {{(pid=61857) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1363.267272] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquiring lock "refresh_cache-6579d6ee-28c8-4c2c-948b-f1b0b8fef482" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1363.267343] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquired lock "refresh_cache-6579d6ee-28c8-4c2c-948b-f1b0b8fef482" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1363.267471] env[61857]: DEBUG nova.network.neutron [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Building network info cache for instance {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1363.799287] env[61857]: DEBUG nova.network.neutron [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Instance cache missing network info. {{(pid=61857) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1363.918154] env[61857]: DEBUG nova.network.neutron [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Updating instance_info_cache with network_info: [{"id": "f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870", "address": "fa:16:3e:0e:c9:3f", "network": {"id": "2f875f13-ba67-4ec2-a193-18ac7f723261", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-65279011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3f950eeb3b468ca7cf97066329f564", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a1439ce-fa5c-420d-bcf0-083f4cc002cd", "external-id": "nsx-vlan-transportzone-413", "segmentation_id": 413, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5a4267f-fa", "ovs_interfaceid": "f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1364.420858] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Releasing lock "refresh_cache-6579d6ee-28c8-4c2c-948b-f1b0b8fef482" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1364.421212] env[61857]: DEBUG nova.compute.manager [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Instance network_info: |[{"id": "f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870", "address": "fa:16:3e:0e:c9:3f", "network": {"id": "2f875f13-ba67-4ec2-a193-18ac7f723261", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-65279011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3f950eeb3b468ca7cf97066329f564", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a1439ce-fa5c-420d-bcf0-083f4cc002cd", "external-id": "nsx-vlan-transportzone-413", "segmentation_id": 413, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5a4267f-fa", "ovs_interfaceid": "f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61857) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1364.421646] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:c9:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0a1439ce-fa5c-420d-bcf0-083f4cc002cd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870', 'vif_model': 'vmxnet3'}] {{(pid=61857) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1364.429054] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Creating folder: Project (6b3f950eeb3b468ca7cf97066329f564). Parent ref: group-v214027. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1364.429380] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a07dce6-aee0-48e3-923d-7a2e284140bc {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.443007] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Created folder: Project (6b3f950eeb3b468ca7cf97066329f564) in parent group-v214027. [ 1364.443202] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Creating folder: Instances. Parent ref: group-v214243. {{(pid=61857) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1364.443737] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8930971-e1a6-4ed5-9f2d-913c8564eab6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.452428] env[61857]: INFO nova.virt.vmwareapi.vm_util [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Created folder: Instances in parent group-v214243. [ 1364.452649] env[61857]: DEBUG oslo.service.loopingcall [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1364.452835] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Creating VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1364.453039] env[61857]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-39620b5e-e7cc-47a3-ad18-760da0de685e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.470154] env[61857]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1364.470154] env[61857]: value = "task-951770" [ 1364.470154] env[61857]: _type = "Task" [ 1364.470154] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.477285] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951770, 'name': CreateVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.710075] env[61857]: DEBUG nova.compute.manager [req-e1b09e5c-434b-4466-84ba-209403fb9077 req-da221e3d-e8fd-44ab-98f1-f451f9bc2c21 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Received event network-changed-f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1364.710289] env[61857]: DEBUG nova.compute.manager [req-e1b09e5c-434b-4466-84ba-209403fb9077 req-da221e3d-e8fd-44ab-98f1-f451f9bc2c21 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Refreshing instance network info cache due to event network-changed-f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870. {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1364.710580] env[61857]: DEBUG oslo_concurrency.lockutils [req-e1b09e5c-434b-4466-84ba-209403fb9077 req-da221e3d-e8fd-44ab-98f1-f451f9bc2c21 service nova] Acquiring lock "refresh_cache-6579d6ee-28c8-4c2c-948b-f1b0b8fef482" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1364.710745] env[61857]: DEBUG oslo_concurrency.lockutils [req-e1b09e5c-434b-4466-84ba-209403fb9077 req-da221e3d-e8fd-44ab-98f1-f451f9bc2c21 service nova] Acquired lock "refresh_cache-6579d6ee-28c8-4c2c-948b-f1b0b8fef482" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1364.710919] env[61857]: DEBUG nova.network.neutron [req-e1b09e5c-434b-4466-84ba-209403fb9077 req-da221e3d-e8fd-44ab-98f1-f451f9bc2c21 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Refreshing network info cache for port f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870 {{(pid=61857) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1364.980320] env[61857]: DEBUG oslo_vmware.api [-] Task: {'id': task-951770, 'name': CreateVM_Task, 'duration_secs': 0.292012} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.980708] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Created VM on the ESX host {{(pid=61857) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1364.981128] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1364.981309] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1364.981665] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1364.981917] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bfd7559-dbe7-4357-81b7-77cce359bcfb {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.986022] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for the task: (returnval){ [ 1364.986022] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52844364-5f3f-6e1c-4c19-4e99b3ad5e0b" [ 1364.986022] env[61857]: _type = "Task" [ 1364.986022] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.993092] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52844364-5f3f-6e1c-4c19-4e99b3ad5e0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.388693] env[61857]: DEBUG nova.network.neutron [req-e1b09e5c-434b-4466-84ba-209403fb9077 req-da221e3d-e8fd-44ab-98f1-f451f9bc2c21 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Updated VIF entry in instance network info cache for port f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870. {{(pid=61857) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1365.389070] env[61857]: DEBUG nova.network.neutron [req-e1b09e5c-434b-4466-84ba-209403fb9077 req-da221e3d-e8fd-44ab-98f1-f451f9bc2c21 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Updating instance_info_cache with network_info: [{"id": "f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870", "address": "fa:16:3e:0e:c9:3f", "network": {"id": "2f875f13-ba67-4ec2-a193-18ac7f723261", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-65279011-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b3f950eeb3b468ca7cf97066329f564", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a1439ce-fa5c-420d-bcf0-083f4cc002cd", "external-id": "nsx-vlan-transportzone-413", "segmentation_id": 413, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5a4267f-fa", "ovs_interfaceid": "f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1365.496205] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52844364-5f3f-6e1c-4c19-4e99b3ad5e0b, 'name': SearchDatastore_Task, 'duration_secs': 0.009889} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.496478] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1365.496782] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Processing image 2e0c8bb4-0827-486c-98ce-7994555ffc23 {{(pid=61857) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1365.497049] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1365.497211] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1365.497398] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1365.497658] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd333683-fc80-4ca7-9c71-f6d45c6be4b1 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.505552] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61857) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1365.505733] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61857) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1365.506414] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63bf8d23-1b2c-4f5a-bad9-e51eece35f39 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.511061] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for the task: (returnval){ [ 1365.511061] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52f007a2-b651-af87-03c0-a2f79c454bc0" [ 1365.511061] env[61857]: _type = "Task" [ 1365.511061] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.518064] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f007a2-b651-af87-03c0-a2f79c454bc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.891849] env[61857]: DEBUG oslo_concurrency.lockutils [req-e1b09e5c-434b-4466-84ba-209403fb9077 req-da221e3d-e8fd-44ab-98f1-f451f9bc2c21 service nova] Releasing lock "refresh_cache-6579d6ee-28c8-4c2c-948b-f1b0b8fef482" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1366.021985] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52f007a2-b651-af87-03c0-a2f79c454bc0, 'name': SearchDatastore_Task, 'duration_secs': 0.007763} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.022763] env[61857]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e35a8174-b0c7-4357-a95c-ac287097d890 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.027630] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for the task: (returnval){ [ 1366.027630] env[61857]: value = "session[5244fa0a-645a-cc16-995b-7d97c674c811]52d0bc02-3212-2226-07d8-6358e3e9d00b" [ 1366.027630] env[61857]: _type = "Task" [ 1366.027630] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.035261] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52d0bc02-3212-2226-07d8-6358e3e9d00b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.538013] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': session[5244fa0a-645a-cc16-995b-7d97c674c811]52d0bc02-3212-2226-07d8-6358e3e9d00b, 'name': SearchDatastore_Task, 'duration_secs': 0.009453} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.538311] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk" {{(pid=61857) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1366.538572] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 6579d6ee-28c8-4c2c-948b-f1b0b8fef482/6579d6ee-28c8-4c2c-948b-f1b0b8fef482.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1366.538850] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51a9ce72-4b80-4cdf-bf85-2563a352dac8 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.544602] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for the task: (returnval){ [ 1366.544602] env[61857]: value = "task-951771" [ 1366.544602] env[61857]: _type = "Task" [ 1366.544602] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.552792] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951771, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.054027] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951771, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.431343} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.054369] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2e0c8bb4-0827-486c-98ce-7994555ffc23/2e0c8bb4-0827-486c-98ce-7994555ffc23.vmdk to [datastore2] 6579d6ee-28c8-4c2c-948b-f1b0b8fef482/6579d6ee-28c8-4c2c-948b-f1b0b8fef482.vmdk {{(pid=61857) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1367.054573] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Extending root virtual disk to 1048576 {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1367.054820] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f8d573c-5893-4669-8d38-daa08f7af19c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.061156] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for the task: (returnval){ [ 1367.061156] env[61857]: value = "task-951772" [ 1367.061156] env[61857]: _type = "Task" [ 1367.061156] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.068253] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951772, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.570374] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951772, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063183} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.570670] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Extended root virtual disk {{(pid=61857) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1367.571416] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fbb2a8-5c24-4124-bb44-759a4624b219 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.593021] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 6579d6ee-28c8-4c2c-948b-f1b0b8fef482/6579d6ee-28c8-4c2c-948b-f1b0b8fef482.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1367.593235] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ae3643a-f035-49e0-b3fa-8fbf0b224167 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.613372] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for the task: (returnval){ [ 1367.613372] env[61857]: value = "task-951773" [ 1367.613372] env[61857]: _type = "Task" [ 1367.613372] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.620666] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951773, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.124075] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951773, 'name': ReconfigVM_Task, 'duration_secs': 0.277914} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1368.124075] env[61857]: DEBUG nova.virt.vmwareapi.volumeops [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 6579d6ee-28c8-4c2c-948b-f1b0b8fef482/6579d6ee-28c8-4c2c-948b-f1b0b8fef482.vmdk or device None with type sparse {{(pid=61857) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1368.124495] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-86c107ed-89f0-4b90-a5f1-c7570ceab110 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.130843] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for the task: (returnval){ [ 1368.130843] env[61857]: value = "task-951774" [ 1368.130843] env[61857]: _type = "Task" [ 1368.130843] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1368.138131] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951774, 'name': Rename_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.640697] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951774, 'name': Rename_Task, 'duration_secs': 0.14897} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1368.641079] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Powering on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1368.641338] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88076ed0-530a-4f47-94f5-bd8e72f08de6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.647780] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for the task: (returnval){ [ 1368.647780] env[61857]: value = "task-951775" [ 1368.647780] env[61857]: _type = "Task" [ 1368.647780] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1368.655114] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951775, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1369.157603] env[61857]: DEBUG oslo_vmware.api [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951775, 'name': PowerOnVM_Task, 'duration_secs': 0.412229} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1369.158032] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Powered on the VM {{(pid=61857) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1369.158101] env[61857]: INFO nova.compute.manager [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Took 6.62 seconds to spawn the instance on the hypervisor. [ 1369.158263] env[61857]: DEBUG nova.compute.manager [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Checking state {{(pid=61857) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1369.159062] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a84cb8-2c35-4498-bf9d-5433facc4e1e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.678008] env[61857]: INFO nova.compute.manager [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Took 11.28 seconds to build instance. [ 1370.180126] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f58200fe-c837-44a0-9ba3-325f57aef5f2 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.794s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1371.342697] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquiring lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1371.342992] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1371.343230] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquiring lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1371.343422] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1371.343604] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1371.345800] env[61857]: INFO nova.compute.manager [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Terminating instance [ 1371.347491] env[61857]: DEBUG nova.compute.manager [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Start destroying the instance on the hypervisor. {{(pid=61857) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1371.347695] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Destroying instance {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1371.348550] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df060ba0-1b35-4e30-bfc2-e6bdd008af0c {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.356071] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Powering off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1371.356298] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-886da91e-2b11-44ed-9897-732d035d9b67 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.362987] env[61857]: DEBUG oslo_vmware.api [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for the task: (returnval){ [ 1371.362987] env[61857]: value = "task-951776" [ 1371.362987] env[61857]: _type = "Task" [ 1371.362987] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1371.370289] env[61857]: DEBUG oslo_vmware.api [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951776, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1371.873394] env[61857]: DEBUG oslo_vmware.api [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951776, 'name': PowerOffVM_Task, 'duration_secs': 0.191763} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1371.873674] env[61857]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Powered off the VM {{(pid=61857) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1371.873821] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Unregistering the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1371.874094] env[61857]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-089a9826-9869-4f5a-84c6-304532d14035 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.953949] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Unregistered the VM {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1371.954223] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Deleting contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1371.954409] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Deleting the datastore file [datastore2] 6579d6ee-28c8-4c2c-948b-f1b0b8fef482 {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1371.954695] env[61857]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8aa646be-8386-4990-948a-3df5e9ab772b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.961019] env[61857]: DEBUG oslo_vmware.api [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for the task: (returnval){ [ 1371.961019] env[61857]: value = "task-951778" [ 1371.961019] env[61857]: _type = "Task" [ 1371.961019] env[61857]: } to complete. {{(pid=61857) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1371.968770] env[61857]: DEBUG oslo_vmware.api [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951778, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1372.471613] env[61857]: DEBUG oslo_vmware.api [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Task: {'id': task-951778, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152994} completed successfully. {{(pid=61857) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1372.472212] env[61857]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Deleted the datastore file {{(pid=61857) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1372.472396] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Deleted contents of the VM from datastore datastore2 {{(pid=61857) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1372.472624] env[61857]: DEBUG nova.virt.vmwareapi.vmops [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Instance destroyed {{(pid=61857) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1372.472890] env[61857]: INFO nova.compute.manager [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1372.473231] env[61857]: DEBUG oslo.service.loopingcall [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61857) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1372.473500] env[61857]: DEBUG nova.compute.manager [-] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Deallocating network for instance {{(pid=61857) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1372.473656] env[61857]: DEBUG nova.network.neutron [-] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] deallocate_for_instance() {{(pid=61857) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1372.703913] env[61857]: DEBUG nova.compute.manager [req-a0d54ac8-4a7a-43c8-9d19-464989293880 req-bdcde2cc-70b3-46bb-9fc9-616cf51aeb78 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Received event network-vif-deleted-f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870 {{(pid=61857) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1372.704155] env[61857]: INFO nova.compute.manager [req-a0d54ac8-4a7a-43c8-9d19-464989293880 req-bdcde2cc-70b3-46bb-9fc9-616cf51aeb78 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Neutron deleted interface f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870; detaching it from the instance and deleting it from the info cache [ 1372.704319] env[61857]: DEBUG nova.network.neutron [req-a0d54ac8-4a7a-43c8-9d19-464989293880 req-bdcde2cc-70b3-46bb-9fc9-616cf51aeb78 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1373.181259] env[61857]: DEBUG nova.network.neutron [-] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Updating instance_info_cache with network_info: [] {{(pid=61857) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1373.206236] env[61857]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae429917-ed5f-473b-b838-00564bcc3901 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.215937] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6357a07-4bda-4b31-9e04-0841a8bfc1b6 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.237348] env[61857]: DEBUG nova.compute.manager [req-a0d54ac8-4a7a-43c8-9d19-464989293880 req-bdcde2cc-70b3-46bb-9fc9-616cf51aeb78 service nova] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Detach interface failed, port_id=f5a4267f-fa7c-4c2c-84ed-c1e2a46e5870, reason: Instance 6579d6ee-28c8-4c2c-948b-f1b0b8fef482 could not be found. {{(pid=61857) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1373.683664] env[61857]: INFO nova.compute.manager [-] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Took 1.21 seconds to deallocate network for instance. [ 1374.190542] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1374.190860] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1374.191099] env[61857]: DEBUG nova.objects.instance [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lazy-loading 'resources' on Instance uuid 6579d6ee-28c8-4c2c-948b-f1b0b8fef482 {{(pid=61857) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1374.723612] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c11bf8-f07d-4efa-98ba-2c903a2ddb2e {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.732671] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ed7b04-5a57-4f70-a6c2-ebd58421cf64 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.761260] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8c439d-b0a6-40d4-9aca-1facdea345f5 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.768481] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4202abf-0210-4273-a1cb-a2482e3de4f9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.780847] env[61857]: DEBUG nova.compute.provider_tree [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1375.284304] env[61857]: DEBUG nova.scheduler.client.report [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1375.789534] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1375.810344] env[61857]: INFO nova.scheduler.client.report [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Deleted allocations for instance 6579d6ee-28c8-4c2c-948b-f1b0b8fef482 [ 1376.318610] env[61857]: DEBUG oslo_concurrency.lockutils [None req-bf3fd723-8959-48b3-87ba-5c5ba2f27149 tempest-ServerMetadataTestJSON-457805838 tempest-ServerMetadataTestJSON-457805838-project-member] Lock "6579d6ee-28c8-4c2c-948b-f1b0b8fef482" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.975s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1380.690138] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1384.689659] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1386.187867] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1386.188306] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.689597] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.689925] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.690088] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Cleaning up deleted instances with incomplete migration {{(pid=61857) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1389.191913] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.192307] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Starting heal instance info cache {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1389.192307] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Rebuilding the list of instances to heal {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1389.695031] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Didn't find any instances for network info cache update. {{(pid=61857) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1389.695320] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.695487] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1391.689743] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1391.690193] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61857) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1392.689323] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1393.192598] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1393.192993] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1393.193120] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1393.193209] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61857) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1393.194533] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409bdc8d-0459-4343-9e62-c2d8dea9eea9 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.202259] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a708aaad-522d-4680-bedf-0d7d2b915846 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.220739] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed242fb0-b7d8-4078-b43b-ad42ebce5274 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.234072] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c512ae79-60d7-4afe-b5e1-d7df03c7f292 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.262649] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181462MB free_disk=41GB free_vcpus=48 pci_devices=None {{(pid=61857) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1393.262759] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1393.262941] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.398919] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1394.399228] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61857) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1394.412045] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195d0e1a-bf69-4eee-972b-e941dbf15031 {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.419581] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e7eb93-a5ee-4e0a-9d09-428d1fa88a5b {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.447980] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6c9f5e-cfca-4b05-8b5b-8712cf07440d {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.454406] env[61857]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c72a752-192c-4eb4-9254-f5cbb18b2e0f {{(pid=61857) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.466822] env[61857]: DEBUG nova.compute.provider_tree [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed in ProviderTree for provider: 2d5860d0-8574-4e55-9ced-20e66f0314c2 {{(pid=61857) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1394.970368] env[61857]: DEBUG nova.scheduler.client.report [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Inventory has not changed for provider 2d5860d0-8574-4e55-9ced-20e66f0314c2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 41, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61857) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1395.474903] env[61857]: DEBUG nova.compute.resource_tracker [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61857) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1395.475225] env[61857]: DEBUG oslo_concurrency.lockutils [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.212s {{(pid=61857) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1395.689608] env[61857]: DEBUG oslo_service.periodic_task [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61857) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1395.689788] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] Cleaning up deleted instances {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1396.195931] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] There are 13 instances to clean {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1396.196307] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 6579d6ee-28c8-4c2c-948b-f1b0b8fef482] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1396.699321] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c8acbb20-a33f-47c3-a37f-d350e1e21880] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1397.202643] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 299da51d-22a9-4da5-992a-482b63b81034] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1397.705819] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: fa2b18b4-b5ba-4036-b50e-ccdefd1f0524] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1398.208749] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: b33cfca7-5621-47a8-bb84-59d82847d699] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1398.712673] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: 857977a6-b81b-4c94-b823-436e57b34ba0] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1399.216460] env[61857]: DEBUG nova.compute.manager [None req-f334939a-cb88-4052-b8aa-637913a6be93 None None] [instance: c6d20166-7251-406f-9003-b0599c2f5987] Instance has had 0 of 5 cleanup attempts {{(pid=61857) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}}